var/home/core/zuul-output/0000755000175000017500000000000015113753616014535 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015113765152015477 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005115170415113765144017710 0ustar rootrootDec 03 06:27:21 crc systemd[1]: Starting Kubernetes Kubelet... Dec 03 06:27:21 crc restorecon[4690]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:21 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:27:22 crc restorecon[4690]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 06:27:22 crc restorecon[4690]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 03 06:27:22 crc kubenswrapper[4818]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 06:27:22 crc kubenswrapper[4818]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 03 06:27:22 crc kubenswrapper[4818]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 06:27:22 crc kubenswrapper[4818]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 06:27:22 crc kubenswrapper[4818]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 03 06:27:22 crc kubenswrapper[4818]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.545436 4818 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557784 4818 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557852 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557859 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557866 4818 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557873 4818 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557879 4818 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557886 4818 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557893 4818 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557899 4818 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557906 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557913 4818 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557918 4818 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557925 4818 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557930 4818 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557936 4818 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557942 4818 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557947 4818 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557953 4818 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557959 4818 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557964 4818 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557970 4818 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557975 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557981 4818 feature_gate.go:330] unrecognized feature gate: Example Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557987 4818 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557992 4818 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.557997 4818 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558003 4818 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558008 4818 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558014 4818 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558019 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558025 4818 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558030 4818 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558035 4818 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558044 4818 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558052 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558060 4818 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558068 4818 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558074 4818 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558081 4818 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558087 4818 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558093 4818 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558098 4818 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558103 4818 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558109 4818 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558114 4818 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558119 4818 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558125 4818 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558130 4818 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558135 4818 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558143 4818 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558150 4818 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558155 4818 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558160 4818 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558166 4818 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558171 4818 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558178 4818 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558186 4818 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558202 4818 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558214 4818 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558224 4818 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558233 4818 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558240 4818 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558248 4818 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558257 4818 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558264 4818 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558270 4818 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558277 4818 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558283 4818 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558289 4818 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558296 4818 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.558302 4818 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558415 4818 flags.go:64] FLAG: --address="0.0.0.0" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558427 4818 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558462 4818 flags.go:64] FLAG: --anonymous-auth="true" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558477 4818 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558485 4818 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558492 4818 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558500 4818 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558508 4818 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558515 4818 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558521 4818 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558529 4818 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558535 4818 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558542 4818 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558548 4818 flags.go:64] FLAG: --cgroup-root="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558554 4818 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558560 4818 flags.go:64] FLAG: --client-ca-file="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558567 4818 flags.go:64] FLAG: --cloud-config="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558574 4818 flags.go:64] FLAG: --cloud-provider="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558581 4818 flags.go:64] FLAG: --cluster-dns="[]" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558593 4818 flags.go:64] FLAG: --cluster-domain="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558600 4818 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558609 4818 flags.go:64] FLAG: --config-dir="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558616 4818 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558624 4818 flags.go:64] FLAG: --container-log-max-files="5" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558633 4818 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558642 4818 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558650 4818 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558695 4818 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558703 4818 flags.go:64] FLAG: --contention-profiling="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558711 4818 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558719 4818 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558727 4818 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558735 4818 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558745 4818 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558754 4818 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558762 4818 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558770 4818 flags.go:64] FLAG: --enable-load-reader="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558778 4818 flags.go:64] FLAG: --enable-server="true" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558786 4818 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558796 4818 flags.go:64] FLAG: --event-burst="100" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558804 4818 flags.go:64] FLAG: --event-qps="50" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558811 4818 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558839 4818 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558846 4818 flags.go:64] FLAG: --eviction-hard="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558856 4818 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558863 4818 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558870 4818 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558881 4818 flags.go:64] FLAG: --eviction-soft="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558888 4818 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558896 4818 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558903 4818 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558911 4818 flags.go:64] FLAG: --experimental-mounter-path="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558918 4818 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558925 4818 flags.go:64] FLAG: --fail-swap-on="true" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558932 4818 flags.go:64] FLAG: --feature-gates="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558941 4818 flags.go:64] FLAG: --file-check-frequency="20s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558948 4818 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558955 4818 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558962 4818 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558969 4818 flags.go:64] FLAG: --healthz-port="10248" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558976 4818 flags.go:64] FLAG: --help="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558983 4818 flags.go:64] FLAG: --hostname-override="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558989 4818 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.558997 4818 flags.go:64] FLAG: --http-check-frequency="20s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559004 4818 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559012 4818 flags.go:64] FLAG: --image-credential-provider-config="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559020 4818 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559028 4818 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559035 4818 flags.go:64] FLAG: --image-service-endpoint="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559043 4818 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559051 4818 flags.go:64] FLAG: --kube-api-burst="100" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559058 4818 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559066 4818 flags.go:64] FLAG: --kube-api-qps="50" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559074 4818 flags.go:64] FLAG: --kube-reserved="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559083 4818 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559091 4818 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559100 4818 flags.go:64] FLAG: --kubelet-cgroups="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559108 4818 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559116 4818 flags.go:64] FLAG: --lock-file="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559123 4818 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559130 4818 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559137 4818 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559146 4818 flags.go:64] FLAG: --log-json-split-stream="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559154 4818 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559160 4818 flags.go:64] FLAG: --log-text-split-stream="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559166 4818 flags.go:64] FLAG: --logging-format="text" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559198 4818 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559207 4818 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559215 4818 flags.go:64] FLAG: --manifest-url="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559223 4818 flags.go:64] FLAG: --manifest-url-header="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559234 4818 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559241 4818 flags.go:64] FLAG: --max-open-files="1000000" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559251 4818 flags.go:64] FLAG: --max-pods="110" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559259 4818 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559268 4818 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559276 4818 flags.go:64] FLAG: --memory-manager-policy="None" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559283 4818 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559292 4818 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559301 4818 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559309 4818 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559329 4818 flags.go:64] FLAG: --node-status-max-images="50" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559335 4818 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559342 4818 flags.go:64] FLAG: --oom-score-adj="-999" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559348 4818 flags.go:64] FLAG: --pod-cidr="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559355 4818 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559364 4818 flags.go:64] FLAG: --pod-manifest-path="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559370 4818 flags.go:64] FLAG: --pod-max-pids="-1" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559376 4818 flags.go:64] FLAG: --pods-per-core="0" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559383 4818 flags.go:64] FLAG: --port="10250" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559389 4818 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559396 4818 flags.go:64] FLAG: --provider-id="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559402 4818 flags.go:64] FLAG: --qos-reserved="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559408 4818 flags.go:64] FLAG: --read-only-port="10255" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559414 4818 flags.go:64] FLAG: --register-node="true" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559421 4818 flags.go:64] FLAG: --register-schedulable="true" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559427 4818 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559447 4818 flags.go:64] FLAG: --registry-burst="10" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559453 4818 flags.go:64] FLAG: --registry-qps="5" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559459 4818 flags.go:64] FLAG: --reserved-cpus="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559470 4818 flags.go:64] FLAG: --reserved-memory="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559478 4818 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559485 4818 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559491 4818 flags.go:64] FLAG: --rotate-certificates="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559497 4818 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559503 4818 flags.go:64] FLAG: --runonce="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559510 4818 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559516 4818 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559523 4818 flags.go:64] FLAG: --seccomp-default="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559529 4818 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559535 4818 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559542 4818 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559548 4818 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559554 4818 flags.go:64] FLAG: --storage-driver-password="root" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559561 4818 flags.go:64] FLAG: --storage-driver-secure="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559567 4818 flags.go:64] FLAG: --storage-driver-table="stats" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559573 4818 flags.go:64] FLAG: --storage-driver-user="root" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559580 4818 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559587 4818 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559593 4818 flags.go:64] FLAG: --system-cgroups="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559599 4818 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559610 4818 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559615 4818 flags.go:64] FLAG: --tls-cert-file="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559621 4818 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559631 4818 flags.go:64] FLAG: --tls-min-version="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559637 4818 flags.go:64] FLAG: --tls-private-key-file="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559643 4818 flags.go:64] FLAG: --topology-manager-policy="none" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559650 4818 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559656 4818 flags.go:64] FLAG: --topology-manager-scope="container" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559663 4818 flags.go:64] FLAG: --v="2" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559670 4818 flags.go:64] FLAG: --version="false" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559678 4818 flags.go:64] FLAG: --vmodule="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559687 4818 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.559693 4818 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559863 4818 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559870 4818 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559878 4818 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559885 4818 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559892 4818 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559898 4818 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559904 4818 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559911 4818 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559916 4818 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559922 4818 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559928 4818 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559933 4818 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559939 4818 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559944 4818 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559950 4818 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559957 4818 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559962 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559969 4818 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559975 4818 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559980 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559986 4818 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559991 4818 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.559996 4818 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560001 4818 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560007 4818 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560022 4818 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560029 4818 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560035 4818 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560041 4818 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560047 4818 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560053 4818 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560059 4818 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560064 4818 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560070 4818 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560076 4818 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560081 4818 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560086 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560091 4818 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560097 4818 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560104 4818 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560111 4818 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560117 4818 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560122 4818 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560128 4818 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560133 4818 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560139 4818 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560144 4818 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560149 4818 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560154 4818 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560159 4818 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560165 4818 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560170 4818 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560175 4818 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560180 4818 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560185 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560190 4818 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560196 4818 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560203 4818 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560209 4818 feature_gate.go:330] unrecognized feature gate: Example Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560214 4818 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560220 4818 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560225 4818 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560230 4818 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560236 4818 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560241 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560246 4818 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560251 4818 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560256 4818 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560261 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560266 4818 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.560272 4818 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.560290 4818 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.569665 4818 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.569714 4818 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.569889 4818 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.569901 4818 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.569911 4818 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.569919 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.569928 4818 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.569936 4818 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.569945 4818 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.569954 4818 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.569963 4818 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.569975 4818 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.569986 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.569995 4818 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570004 4818 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570012 4818 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570020 4818 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570028 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570036 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570045 4818 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570052 4818 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570060 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570068 4818 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570076 4818 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570083 4818 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570091 4818 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570099 4818 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570108 4818 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570116 4818 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570123 4818 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570131 4818 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570139 4818 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570147 4818 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570155 4818 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570163 4818 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570171 4818 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570181 4818 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570189 4818 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570197 4818 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570205 4818 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570213 4818 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570220 4818 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570230 4818 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570238 4818 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570246 4818 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570254 4818 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570262 4818 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570270 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570279 4818 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570287 4818 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570297 4818 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570307 4818 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570316 4818 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570325 4818 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570333 4818 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570344 4818 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570353 4818 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570362 4818 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570370 4818 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570379 4818 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570387 4818 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570395 4818 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570403 4818 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570411 4818 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570421 4818 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570431 4818 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570439 4818 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570446 4818 feature_gate.go:330] unrecognized feature gate: Example Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570454 4818 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570462 4818 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570470 4818 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570479 4818 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570488 4818 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.570501 4818 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570722 4818 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570794 4818 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570803 4818 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570813 4818 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570844 4818 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570853 4818 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570862 4818 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570871 4818 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570879 4818 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570886 4818 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570895 4818 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570903 4818 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570911 4818 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570919 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570959 4818 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570970 4818 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570980 4818 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570989 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.570998 4818 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571006 4818 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571014 4818 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571022 4818 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571029 4818 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571037 4818 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571048 4818 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571058 4818 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571067 4818 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571076 4818 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571085 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571093 4818 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571103 4818 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571113 4818 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571121 4818 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571129 4818 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571149 4818 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571157 4818 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571165 4818 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571173 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571181 4818 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571189 4818 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571197 4818 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571205 4818 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571213 4818 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571223 4818 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571233 4818 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571241 4818 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571249 4818 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571257 4818 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571265 4818 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571274 4818 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571281 4818 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571290 4818 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571298 4818 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571306 4818 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571313 4818 feature_gate.go:330] unrecognized feature gate: Example Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571321 4818 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571329 4818 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571337 4818 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571345 4818 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571352 4818 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571361 4818 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571368 4818 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571377 4818 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571385 4818 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571393 4818 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571400 4818 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571408 4818 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571416 4818 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571424 4818 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571431 4818 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.571440 4818 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.571452 4818 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.571702 4818 server.go:940] "Client rotation is on, will bootstrap in background" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.575762 4818 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.575926 4818 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.576671 4818 server.go:997] "Starting client certificate rotation" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.576715 4818 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.576989 4818 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-18 01:09:27.921327582 +0000 UTC Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.577131 4818 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 354h42m5.344199522s for next certificate rotation Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.588907 4818 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.597585 4818 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.615156 4818 log.go:25] "Validated CRI v1 runtime API" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.628661 4818 log.go:25] "Validated CRI v1 image API" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.630383 4818 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.638789 4818 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-03-06-22-49-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.638840 4818 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.652665 4818 manager.go:217] Machine: {Timestamp:2025-12-03 06:27:22.651372343 +0000 UTC m=+0.342981115 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:af2f867d-be4e-4cab-a642-bf5ce3ee5141 BootID:b45b9166-fc05-4a5d-ae82-e08efaed5d6b Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:64:ed:2d Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:64:ed:2d Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:f4:a6:52 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:5f:0a:f4 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:58:cf:db Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:1f:26:ef Speed:-1 Mtu:1496} {Name:eth10 MacAddress:3a:f3:87:8c:d8:55 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:ae:c6:2e:f6:2d:b5 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.652941 4818 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.653086 4818 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.653781 4818 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.654082 4818 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.654126 4818 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.654378 4818 topology_manager.go:138] "Creating topology manager with none policy" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.654398 4818 container_manager_linux.go:303] "Creating device plugin manager" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.654694 4818 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.654747 4818 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.655308 4818 state_mem.go:36] "Initialized new in-memory state store" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.656082 4818 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.656773 4818 kubelet.go:418] "Attempting to sync node with API server" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.656798 4818 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.656846 4818 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.656865 4818 kubelet.go:324] "Adding apiserver pod source" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.656883 4818 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.658992 4818 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.659561 4818 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.660327 4818 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.660930 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.660963 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.660977 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.660991 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.661016 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.661036 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.661055 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.661084 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.661104 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.661118 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.661138 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.661165 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.661467 4818 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.662384 4818 server.go:1280] "Started kubelet" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.662545 4818 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.662701 4818 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.667467 4818 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 03 06:27:22 crc systemd[1]: Started Kubernetes Kubelet. Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.668313 4818 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.668495 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Dec 03 06:27:22 crc kubenswrapper[4818]: E1203 06:27:22.668621 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.670038 4818 server.go:460] "Adding debug handlers to kubelet server" Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.669470 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Dec 03 06:27:22 crc kubenswrapper[4818]: E1203 06:27:22.670493 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:27:22 crc kubenswrapper[4818]: E1203 06:27:22.677537 4818 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.75:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187da0a112a86209 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 06:27:22.662306313 +0000 UTC m=+0.353915125,LastTimestamp:2025-12-03 06:27:22.662306313 +0000 UTC m=+0.353915125,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.679325 4818 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.679420 4818 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.679474 4818 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 12:14:55.107492995 +0000 UTC Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.679518 4818 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 461h47m32.427977957s for next certificate rotation Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.679533 4818 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.679564 4818 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.679717 4818 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 03 06:27:22 crc kubenswrapper[4818]: E1203 06:27:22.679558 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 06:27:22 crc kubenswrapper[4818]: E1203 06:27:22.679987 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="200ms" Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.680057 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Dec 03 06:27:22 crc kubenswrapper[4818]: E1203 06:27:22.680140 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.680372 4818 factory.go:55] Registering systemd factory Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.680405 4818 factory.go:221] Registration of the systemd container factory successfully Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.681115 4818 factory.go:153] Registering CRI-O factory Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.681157 4818 factory.go:221] Registration of the crio container factory successfully Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.683287 4818 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.683357 4818 factory.go:103] Registering Raw factory Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.683384 4818 manager.go:1196] Started watching for new ooms in manager Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.684238 4818 manager.go:319] Starting recovery of all containers Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.696974 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697096 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697134 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697177 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697211 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697248 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697280 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697311 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697355 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697388 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697430 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697458 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697496 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697540 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697581 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697608 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697635 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697763 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697795 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697817 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697855 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697870 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697885 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697904 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697918 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697936 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697959 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.697986 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698004 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698023 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698041 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698061 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698080 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698099 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698119 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698137 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698158 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698173 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698189 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698208 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698225 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698246 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698261 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698277 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698298 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698313 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698335 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.698392 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699241 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699262 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699279 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699294 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699316 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699330 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699345 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699360 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699373 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699387 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699401 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699417 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699430 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699444 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699457 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699472 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699486 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699535 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699552 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699568 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699587 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699601 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699616 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699633 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699649 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699665 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699683 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699700 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699716 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699737 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699752 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699768 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699785 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699802 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699840 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699854 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699867 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699882 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699897 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699912 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699926 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699942 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699958 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699973 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.699988 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700003 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700019 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700034 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700054 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700110 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700127 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700142 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700155 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700168 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700185 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700200 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700227 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700244 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700259 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700274 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700291 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700307 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700323 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700341 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700357 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700371 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700388 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700402 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700414 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700427 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700908 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700927 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700943 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700958 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700974 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.700992 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.701008 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.701023 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.701041 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.701055 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.701071 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.701085 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.701100 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.701115 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.701130 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.701144 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.701158 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.701173 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.701188 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.701204 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.701218 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.701233 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.701986 4818 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702018 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702037 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702052 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702068 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702083 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702104 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702118 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702131 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702147 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702163 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702176 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702192 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702205 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702217 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702232 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702244 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702259 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702273 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702287 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702301 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702314 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702331 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702345 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702359 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702371 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702384 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702397 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702410 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702420 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702441 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702452 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702520 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702534 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702547 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702559 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702574 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702587 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702603 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702619 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702632 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702646 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702660 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702675 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702688 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702699 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702712 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702725 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702740 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702754 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702768 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702780 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702792 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702806 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702846 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702861 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702876 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702889 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702903 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702919 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702935 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702952 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702969 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702984 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.702998 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.703013 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.703027 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.703041 4818 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.703054 4818 reconstruct.go:97] "Volume reconstruction finished" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.703061 4818 reconciler.go:26] "Reconciler: start to sync state" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.708915 4818 manager.go:324] Recovery completed Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.721168 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.723004 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.723049 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.723065 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.723855 4818 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.723878 4818 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.723904 4818 state_mem.go:36] "Initialized new in-memory state store" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.732504 4818 policy_none.go:49] "None policy: Start" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.733408 4818 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.735452 4818 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.735490 4818 state_mem.go:35] "Initializing new in-memory state store" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.736221 4818 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.736279 4818 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.736319 4818 kubelet.go:2335] "Starting kubelet main sync loop" Dec 03 06:27:22 crc kubenswrapper[4818]: E1203 06:27:22.736383 4818 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 03 06:27:22 crc kubenswrapper[4818]: W1203 06:27:22.738103 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Dec 03 06:27:22 crc kubenswrapper[4818]: E1203 06:27:22.738216 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:27:22 crc kubenswrapper[4818]: E1203 06:27:22.779866 4818 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.807056 4818 manager.go:334] "Starting Device Plugin manager" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.807159 4818 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.807180 4818 server.go:79] "Starting device plugin registration server" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.807797 4818 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.807899 4818 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.808023 4818 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.808224 4818 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.808234 4818 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 03 06:27:22 crc kubenswrapper[4818]: E1203 06:27:22.816912 4818 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.837255 4818 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.837333 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.838533 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.838660 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.838711 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.839002 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.839214 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.839278 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.840414 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.840451 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.840462 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.841035 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.841085 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.841097 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.841347 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.841357 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.841386 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.842330 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.842363 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.842378 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.842441 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.842519 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.842544 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.843616 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.844442 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.844560 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.845933 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.845969 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.845986 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.846072 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.846123 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.846152 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.846194 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.846479 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.847054 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.847523 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.847556 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.847567 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.847756 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.847791 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.847787 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.847869 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.847888 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.848489 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.848526 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.848538 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:22 crc kubenswrapper[4818]: E1203 06:27:22.880751 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="400ms" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.904650 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.904690 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.904728 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.904877 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.904943 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.904996 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.905037 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.905072 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.905093 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.905120 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.905144 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.905259 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.905318 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.905366 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.905391 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.908318 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.909654 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.909681 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.909689 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:22 crc kubenswrapper[4818]: I1203 06:27:22.909710 4818 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 06:27:22 crc kubenswrapper[4818]: E1203 06:27:22.910300 4818 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.75:6443: connect: connection refused" node="crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006256 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006312 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006335 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006354 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006371 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006448 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006547 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006468 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006511 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006520 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006558 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006471 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006585 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006562 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006697 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006742 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006769 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006785 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006825 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006801 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006850 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006885 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006918 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006951 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006982 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.007008 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.007043 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006958 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.007085 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.006984 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.111381 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.112542 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.112589 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.112599 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.112625 4818 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 06:27:23 crc kubenswrapper[4818]: E1203 06:27:23.113088 4818 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.75:6443: connect: connection refused" node="crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.175307 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: W1203 06:27:23.195754 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-f7ccf248ecc9081bb231faac67462aa1ad01526fc942c66364cdb54068535909 WatchSource:0}: Error finding container f7ccf248ecc9081bb231faac67462aa1ad01526fc942c66364cdb54068535909: Status 404 returned error can't find the container with id f7ccf248ecc9081bb231faac67462aa1ad01526fc942c66364cdb54068535909 Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.199044 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.212611 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: W1203 06:27:23.226651 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-36e67c697e1b98cb98230e3c4fc3a56866d36c15b969d4ca533fcd7698cdcf60 WatchSource:0}: Error finding container 36e67c697e1b98cb98230e3c4fc3a56866d36c15b969d4ca533fcd7698cdcf60: Status 404 returned error can't find the container with id 36e67c697e1b98cb98230e3c4fc3a56866d36c15b969d4ca533fcd7698cdcf60 Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.237370 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.245410 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:27:23 crc kubenswrapper[4818]: W1203 06:27:23.260772 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-eb08d84af58516302feb12b9002bf1f838da757e23ac1ae9d8f5a4fa45a1f2be WatchSource:0}: Error finding container eb08d84af58516302feb12b9002bf1f838da757e23ac1ae9d8f5a4fa45a1f2be: Status 404 returned error can't find the container with id eb08d84af58516302feb12b9002bf1f838da757e23ac1ae9d8f5a4fa45a1f2be Dec 03 06:27:23 crc kubenswrapper[4818]: W1203 06:27:23.263563 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-38754c9b2965157f47d7a34f976e26050e4ce59a5fcaa515fbc68743eaaef8aa WatchSource:0}: Error finding container 38754c9b2965157f47d7a34f976e26050e4ce59a5fcaa515fbc68743eaaef8aa: Status 404 returned error can't find the container with id 38754c9b2965157f47d7a34f976e26050e4ce59a5fcaa515fbc68743eaaef8aa Dec 03 06:27:23 crc kubenswrapper[4818]: E1203 06:27:23.281406 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="800ms" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.513201 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.524905 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.524978 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.525004 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.525052 4818 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 06:27:23 crc kubenswrapper[4818]: E1203 06:27:23.525854 4818 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.75:6443: connect: connection refused" node="crc" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.669403 4818 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.741887 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86"} Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.741992 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"38754c9b2965157f47d7a34f976e26050e4ce59a5fcaa515fbc68743eaaef8aa"} Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.743154 4818 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5" exitCode=0 Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.743212 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5"} Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.743287 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"eb08d84af58516302feb12b9002bf1f838da757e23ac1ae9d8f5a4fa45a1f2be"} Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.743380 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.744128 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.744161 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.744176 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.745367 4818 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c" exitCode=0 Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.745405 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c"} Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.745455 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"36e67c697e1b98cb98230e3c4fc3a56866d36c15b969d4ca533fcd7698cdcf60"} Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.745550 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.746456 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.747369 4818 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="af950b9c18ceb135728e96c4407545a21310fba3b6579fe19440d8e234d1bfdf" exitCode=0 Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.747435 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.747460 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.747453 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"af950b9c18ceb135728e96c4407545a21310fba3b6579fe19440d8e234d1bfdf"} Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.747469 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.747435 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.747569 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.747573 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.747677 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.747495 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"2c07f9f5c5068f8e0307893fa414a5e1cd775f10e8c5271b0e73cc398c358d0c"} Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.748213 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.748230 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.748242 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.749630 4818 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa" exitCode=0 Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.749657 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa"} Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.749674 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f7ccf248ecc9081bb231faac67462aa1ad01526fc942c66364cdb54068535909"} Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.750044 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.751074 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.751114 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:23 crc kubenswrapper[4818]: I1203 06:27:23.751127 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:24 crc kubenswrapper[4818]: W1203 06:27:24.065405 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Dec 03 06:27:24 crc kubenswrapper[4818]: E1203 06:27:24.065491 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:27:24 crc kubenswrapper[4818]: E1203 06:27:24.082532 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="1.6s" Dec 03 06:27:24 crc kubenswrapper[4818]: W1203 06:27:24.164697 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Dec 03 06:27:24 crc kubenswrapper[4818]: E1203 06:27:24.165554 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:27:24 crc kubenswrapper[4818]: W1203 06:27:24.181608 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Dec 03 06:27:24 crc kubenswrapper[4818]: E1203 06:27:24.181682 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:27:24 crc kubenswrapper[4818]: W1203 06:27:24.256520 4818 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.75:6443: connect: connection refused Dec 03 06:27:24 crc kubenswrapper[4818]: E1203 06:27:24.256587 4818 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.75:6443: connect: connection refused" logger="UnhandledError" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.326619 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.328292 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.328349 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.328360 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.328392 4818 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 06:27:24 crc kubenswrapper[4818]: E1203 06:27:24.328946 4818 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.75:6443: connect: connection refused" node="crc" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.756978 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5"} Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.757029 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184"} Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.757041 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6"} Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.757051 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d"} Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.757059 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a"} Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.757077 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.758048 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.758078 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.758088 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.758718 4818 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16" exitCode=0 Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.758782 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.758868 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16"} Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.759365 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.759388 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.759398 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.761247 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.761254 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"b87293b3170ab7adc2e618aa0ad71d450908890732c63a165d2b3632eb8a57c5"} Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.762081 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.762142 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.762172 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.763709 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6a192429ab94e49222181fd50bb4e9597966a1b6adb38c8ebaf0eb31fcf6b8bf"} Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.763744 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5faa1cad79099f32db2b2895d0fb547031ee6dfa6c4a9e306f355b93e4c7c959"} Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.763757 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8047009041ecc11795cad4ced8b06bc2a4810c42023810db409854dfcaae2275"} Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.763904 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.764845 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.764869 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.764878 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.771526 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079"} Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.771555 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7"} Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.771564 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783"} Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.771627 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.772241 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.772259 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:24 crc kubenswrapper[4818]: I1203 06:27:24.772267 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.227303 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.235800 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.363399 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.776506 4818 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e" exitCode=0 Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.776609 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.776651 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.776683 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.776700 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.776959 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e"} Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.777080 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.777086 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.777143 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.778038 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.778081 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.778099 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.778633 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.778662 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.778674 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.778695 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.778725 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.778725 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.778764 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.778783 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.778851 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.778874 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.778890 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.778744 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.929187 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.930324 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.930380 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.930400 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:25 crc kubenswrapper[4818]: I1203 06:27:25.930438 4818 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 06:27:26 crc kubenswrapper[4818]: I1203 06:27:26.785396 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86"} Dec 03 06:27:26 crc kubenswrapper[4818]: I1203 06:27:26.785886 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff"} Dec 03 06:27:26 crc kubenswrapper[4818]: I1203 06:27:26.785921 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4"} Dec 03 06:27:26 crc kubenswrapper[4818]: I1203 06:27:26.785950 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44"} Dec 03 06:27:26 crc kubenswrapper[4818]: I1203 06:27:26.785523 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:27:26 crc kubenswrapper[4818]: I1203 06:27:26.786053 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:26 crc kubenswrapper[4818]: I1203 06:27:26.788575 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:26 crc kubenswrapper[4818]: I1203 06:27:26.788605 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:26 crc kubenswrapper[4818]: I1203 06:27:26.788618 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:26 crc kubenswrapper[4818]: I1203 06:27:26.908891 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:26 crc kubenswrapper[4818]: I1203 06:27:26.909115 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:26 crc kubenswrapper[4818]: I1203 06:27:26.910231 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:26 crc kubenswrapper[4818]: I1203 06:27:26.910268 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:26 crc kubenswrapper[4818]: I1203 06:27:26.910281 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:26 crc kubenswrapper[4818]: I1203 06:27:26.931143 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:27 crc kubenswrapper[4818]: I1203 06:27:27.370296 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:27 crc kubenswrapper[4818]: I1203 06:27:27.794731 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c"} Dec 03 06:27:27 crc kubenswrapper[4818]: I1203 06:27:27.794771 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:27 crc kubenswrapper[4818]: I1203 06:27:27.794852 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:27 crc kubenswrapper[4818]: I1203 06:27:27.795791 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:27 crc kubenswrapper[4818]: I1203 06:27:27.795853 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:27 crc kubenswrapper[4818]: I1203 06:27:27.795872 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:27 crc kubenswrapper[4818]: I1203 06:27:27.796494 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:27 crc kubenswrapper[4818]: I1203 06:27:27.796519 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:27 crc kubenswrapper[4818]: I1203 06:27:27.796528 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:28 crc kubenswrapper[4818]: I1203 06:27:28.364247 4818 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 06:27:28 crc kubenswrapper[4818]: I1203 06:27:28.364340 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 06:27:28 crc kubenswrapper[4818]: I1203 06:27:28.797647 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:28 crc kubenswrapper[4818]: I1203 06:27:28.797663 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:28 crc kubenswrapper[4818]: I1203 06:27:28.799126 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:28 crc kubenswrapper[4818]: I1203 06:27:28.799160 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:28 crc kubenswrapper[4818]: I1203 06:27:28.799197 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:28 crc kubenswrapper[4818]: I1203 06:27:28.799227 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:28 crc kubenswrapper[4818]: I1203 06:27:28.799261 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:28 crc kubenswrapper[4818]: I1203 06:27:28.799280 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:28 crc kubenswrapper[4818]: I1203 06:27:28.902280 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:27:28 crc kubenswrapper[4818]: I1203 06:27:28.902569 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:28 crc kubenswrapper[4818]: I1203 06:27:28.904759 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:28 crc kubenswrapper[4818]: I1203 06:27:28.904804 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:28 crc kubenswrapper[4818]: I1203 06:27:28.904847 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:29 crc kubenswrapper[4818]: I1203 06:27:29.825460 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 03 06:27:29 crc kubenswrapper[4818]: I1203 06:27:29.825703 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:29 crc kubenswrapper[4818]: I1203 06:27:29.826784 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:29 crc kubenswrapper[4818]: I1203 06:27:29.826839 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:29 crc kubenswrapper[4818]: I1203 06:27:29.826849 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:32 crc kubenswrapper[4818]: I1203 06:27:32.104354 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:27:32 crc kubenswrapper[4818]: I1203 06:27:32.104598 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:32 crc kubenswrapper[4818]: I1203 06:27:32.106104 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:32 crc kubenswrapper[4818]: I1203 06:27:32.106171 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:32 crc kubenswrapper[4818]: I1203 06:27:32.106184 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:32 crc kubenswrapper[4818]: I1203 06:27:32.289484 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:27:32 crc kubenswrapper[4818]: I1203 06:27:32.289769 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:32 crc kubenswrapper[4818]: I1203 06:27:32.291385 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:32 crc kubenswrapper[4818]: I1203 06:27:32.291452 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:32 crc kubenswrapper[4818]: I1203 06:27:32.291473 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:32 crc kubenswrapper[4818]: E1203 06:27:32.817119 4818 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 06:27:33 crc kubenswrapper[4818]: I1203 06:27:33.510459 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 03 06:27:33 crc kubenswrapper[4818]: I1203 06:27:33.510757 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:33 crc kubenswrapper[4818]: I1203 06:27:33.512337 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:33 crc kubenswrapper[4818]: I1203 06:27:33.512406 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:33 crc kubenswrapper[4818]: I1203 06:27:33.512421 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:34 crc kubenswrapper[4818]: I1203 06:27:34.670405 4818 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 03 06:27:35 crc kubenswrapper[4818]: I1203 06:27:35.331210 4818 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 06:27:35 crc kubenswrapper[4818]: I1203 06:27:35.331286 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 06:27:35 crc kubenswrapper[4818]: I1203 06:27:35.338632 4818 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 06:27:35 crc kubenswrapper[4818]: I1203 06:27:35.338705 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 06:27:37 crc kubenswrapper[4818]: I1203 06:27:37.376452 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:37 crc kubenswrapper[4818]: I1203 06:27:37.376664 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:37 crc kubenswrapper[4818]: I1203 06:27:37.378201 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:37 crc kubenswrapper[4818]: I1203 06:27:37.378247 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:37 crc kubenswrapper[4818]: I1203 06:27:37.378256 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:37 crc kubenswrapper[4818]: I1203 06:27:37.381854 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:37 crc kubenswrapper[4818]: I1203 06:27:37.822659 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:27:37 crc kubenswrapper[4818]: I1203 06:27:37.822709 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:37 crc kubenswrapper[4818]: I1203 06:27:37.823638 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:37 crc kubenswrapper[4818]: I1203 06:27:37.823694 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:37 crc kubenswrapper[4818]: I1203 06:27:37.823712 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:38 crc kubenswrapper[4818]: I1203 06:27:38.363950 4818 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 06:27:38 crc kubenswrapper[4818]: I1203 06:27:38.364055 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 06:27:38 crc kubenswrapper[4818]: I1203 06:27:38.907636 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:27:38 crc kubenswrapper[4818]: I1203 06:27:38.907813 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:38 crc kubenswrapper[4818]: I1203 06:27:38.909152 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:38 crc kubenswrapper[4818]: I1203 06:27:38.909196 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:38 crc kubenswrapper[4818]: I1203 06:27:38.909206 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.317094 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.330567 4818 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.330809 4818 trace.go:236] Trace[1632600770]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 06:27:26.621) (total time: 13709ms): Dec 03 06:27:40 crc kubenswrapper[4818]: Trace[1632600770]: ---"Objects listed" error: 13709ms (06:27:40.330) Dec 03 06:27:40 crc kubenswrapper[4818]: Trace[1632600770]: [13.709045291s] [13.709045291s] END Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.330900 4818 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.332093 4818 trace.go:236] Trace[764428734]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 06:27:26.124) (total time: 14207ms): Dec 03 06:27:40 crc kubenswrapper[4818]: Trace[764428734]: ---"Objects listed" error: 14207ms (06:27:40.331) Dec 03 06:27:40 crc kubenswrapper[4818]: Trace[764428734]: [14.20780201s] [14.20780201s] END Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.332131 4818 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.339234 4818 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.339568 4818 trace.go:236] Trace[282247887]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 06:27:27.377) (total time: 12962ms): Dec 03 06:27:40 crc kubenswrapper[4818]: Trace[282247887]: ---"Objects listed" error: 12962ms (06:27:40.339) Dec 03 06:27:40 crc kubenswrapper[4818]: Trace[282247887]: [12.962217677s] [12.962217677s] END Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.339605 4818 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.339881 4818 trace.go:236] Trace[2030528194]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 06:27:26.507) (total time: 13832ms): Dec 03 06:27:40 crc kubenswrapper[4818]: Trace[2030528194]: ---"Objects listed" error: 13832ms (06:27:40.339) Dec 03 06:27:40 crc kubenswrapper[4818]: Trace[2030528194]: [13.832649796s] [13.832649796s] END Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.339919 4818 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.551804 4818 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37092->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.551876 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37092->192.168.126.11:17697: read: connection reset by peer" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.551804 4818 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37082->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.552128 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37082->192.168.126.11:17697: read: connection reset by peer" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.552644 4818 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.552706 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.667730 4818 apiserver.go:52] "Watching apiserver" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.669966 4818 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.670202 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.670539 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.671115 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.671256 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.672010 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.672017 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.673060 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.673215 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.673283 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.673371 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.674932 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.675381 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.676407 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.676419 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.676886 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.676976 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.677664 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.677976 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.678040 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.680322 4818 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.701107 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.710445 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.720452 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742033 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742067 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742083 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742100 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742116 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742132 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742173 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742190 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742205 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742219 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742236 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742253 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742267 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742282 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742297 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742310 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742327 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742371 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742387 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742400 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742414 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742455 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742471 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742485 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742498 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742514 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742531 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742545 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742560 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742574 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742594 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742609 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742624 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742641 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742658 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742694 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742709 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742725 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742741 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742755 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742770 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742784 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742800 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742829 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742845 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742859 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742874 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742888 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742903 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742918 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742933 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742948 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742969 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.742982 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.743006 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.744694 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.744791 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.744853 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745232 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745233 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745274 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745314 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745345 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745374 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745395 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745424 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745459 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745484 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745515 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745543 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745571 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745602 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745644 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745691 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745724 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745756 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745780 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745828 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745861 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745893 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745922 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745952 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745985 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746082 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746113 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746151 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746183 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746208 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746237 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746336 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746367 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746398 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746431 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746465 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746492 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746565 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746597 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746626 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746655 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746682 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746704 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746733 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746764 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746796 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746839 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746876 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746906 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746931 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746970 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747003 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747029 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747059 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747090 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747122 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747148 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747243 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747279 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747307 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747341 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747372 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747403 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747432 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747465 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747498 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747524 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747554 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747586 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747612 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747645 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747678 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747708 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747735 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747765 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747792 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747834 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747868 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747898 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747924 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747950 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748050 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748093 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748120 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748144 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748172 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748200 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748224 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748251 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748280 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748309 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748332 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748357 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748383 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748408 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748435 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748466 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748488 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748512 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748537 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748564 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748589 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748621 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748652 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748676 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748702 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748729 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748753 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748778 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748805 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748852 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748878 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748907 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748937 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748961 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748989 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749019 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749046 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749071 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749100 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749130 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749167 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749197 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749228 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749253 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749283 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749313 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749346 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749370 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749396 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749427 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749457 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749486 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749516 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749544 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749569 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749599 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749627 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749678 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749746 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749777 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749808 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749858 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749897 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749927 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749957 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.749990 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.750024 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.750053 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.750079 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.750112 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.750143 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.750174 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.750264 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.750282 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745416 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745447 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.758502 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745460 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745484 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745545 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745560 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745590 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745751 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745760 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745781 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745891 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745932 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.745947 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746009 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746120 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746175 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746243 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746280 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746398 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746395 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746584 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746714 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746741 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746778 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746767 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.746995 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747055 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747129 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747148 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747204 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747239 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747265 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747391 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747400 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747445 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747448 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747466 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747471 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747628 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747836 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747808 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.747965 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748126 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748274 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.748548 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.752650 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.752961 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.753058 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.753520 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.753547 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.754019 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.754203 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.754338 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.754546 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.755271 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.755316 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.755610 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.755683 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.755771 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.755971 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.756162 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.756397 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.756447 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.756746 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.757234 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.757410 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.757491 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.757621 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.757687 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.757896 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.758165 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.758243 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.758312 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.758421 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.758514 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.758797 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.758897 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.758944 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.758983 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.759057 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.759062 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:27:41.259033149 +0000 UTC m=+18.950641921 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.759776 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.759959 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.759805 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.759364 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.759531 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.759728 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.760292 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.760388 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.760602 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.760625 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.760695 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.760964 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.761077 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.761083 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.761311 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.761469 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.761561 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.761662 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.761773 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.761949 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.762019 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.762105 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.762346 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.762373 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.762960 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.763151 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.763263 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.763368 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.763528 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.764000 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.764013 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.764089 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.764112 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.764010 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.764216 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.764336 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.764260 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.764353 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.764374 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.764525 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.764479 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.764955 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.765128 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.765102 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.765333 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.765370 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.764967 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.765456 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.765577 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.765905 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.766193 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.766252 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.766279 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.766301 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.766396 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.766441 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.766523 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.766592 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.766634 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.766784 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.767354 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.767400 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.767622 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.768112 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.768149 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.768398 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.759215 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.768664 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.771441 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.771630 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.771981 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.772098 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.772468 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.773001 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.773651 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.774141 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.774837 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.775305 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.775395 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.775414 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.775435 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.775537 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:41.275515151 +0000 UTC m=+18.967123983 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.775568 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.775630 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.775722 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:41.27568693 +0000 UTC m=+18.967295682 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.775749 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.775895 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.776103 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.776454 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.776421 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.776578 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.776678 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.776749 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.776940 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.776977 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.777008 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.777306 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.777423 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.777608 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.777776 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.778001 4818 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.778024 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.778215 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.778312 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.778499 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.778632 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.778913 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.779051 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.779101 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.779134 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.779494 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.779589 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.779978 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.780664 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.780694 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.780757 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:41.280737779 +0000 UTC m=+18.972346541 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.780341 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.780054 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.780194 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.780601 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.787901 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.788054 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.789806 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.790772 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.790957 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.791774 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.791795 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.791808 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:40 crc kubenswrapper[4818]: E1203 06:27:40.791864 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:41.291849593 +0000 UTC m=+18.983458345 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.792657 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.796514 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.801318 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.801793 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.802093 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.803026 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.810384 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.829483 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.831140 4818 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5" exitCode=255 Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.831201 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5"} Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.841189 4818 scope.go:117] "RemoveContainer" containerID="27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.841542 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.841782 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.851502 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.851590 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.851729 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.851739 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.851776 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.852115 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.852931 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853284 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853311 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853323 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853337 4818 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853348 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853359 4818 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853370 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853448 4818 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853494 4818 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853505 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853514 4818 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853544 4818 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853556 4818 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853566 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853575 4818 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853586 4818 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853600 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853608 4818 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853617 4818 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853626 4818 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853635 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853648 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853659 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853669 4818 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853681 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853690 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853698 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853707 4818 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853716 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853731 4818 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853743 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853752 4818 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853790 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853801 4818 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853810 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853836 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853846 4818 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853856 4818 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853865 4818 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853875 4818 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853885 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853895 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853910 4818 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853920 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853930 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853939 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853949 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853959 4818 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853968 4818 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853979 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.853990 4818 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854001 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854015 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854027 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854042 4818 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854113 4818 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854127 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854139 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854149 4818 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854159 4818 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854169 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854178 4818 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854188 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854198 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854206 4818 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854216 4818 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854225 4818 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854236 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854250 4818 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854260 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854269 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854278 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854288 4818 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854297 4818 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854310 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854320 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854330 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854343 4818 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854353 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854363 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854373 4818 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854382 4818 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854392 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854404 4818 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854421 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854430 4818 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854441 4818 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854451 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854462 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854471 4818 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854482 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854490 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854500 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854508 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854518 4818 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854527 4818 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854536 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854579 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854590 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854601 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854611 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854644 4818 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854654 4818 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854664 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854672 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854683 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854693 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854705 4818 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854717 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854727 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854738 4818 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854746 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854755 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854764 4818 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854773 4818 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854782 4818 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854792 4818 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854802 4818 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854826 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854835 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854843 4818 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854852 4818 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854861 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854869 4818 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854879 4818 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854888 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854896 4818 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854913 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854939 4818 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854947 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854956 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854965 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854974 4818 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854984 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.854993 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855002 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855011 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855024 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855034 4818 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855047 4818 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855057 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855068 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855077 4818 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855088 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855097 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855107 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855116 4818 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855133 4818 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855144 4818 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855153 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855162 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855172 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855180 4818 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855191 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855200 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855216 4818 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855225 4818 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855235 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855244 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855252 4818 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855262 4818 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855271 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855281 4818 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855290 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855299 4818 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855309 4818 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855319 4818 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855331 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855343 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855355 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855368 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855380 4818 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855391 4818 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855401 4818 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855410 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855421 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855431 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855439 4818 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855447 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855457 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855466 4818 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855479 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855490 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855502 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855513 4818 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855524 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855535 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855547 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855559 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.855568 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.862716 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.873987 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.885241 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.899165 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.984628 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 06:27:40 crc kubenswrapper[4818]: I1203 06:27:40.994627 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 06:27:40 crc kubenswrapper[4818]: W1203 06:27:40.999912 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-d7a42366b6cb1d4363e8ed0d5b04fd55a1d4b54074fce899c5525714259439fd WatchSource:0}: Error finding container d7a42366b6cb1d4363e8ed0d5b04fd55a1d4b54074fce899c5525714259439fd: Status 404 returned error can't find the container with id d7a42366b6cb1d4363e8ed0d5b04fd55a1d4b54074fce899c5525714259439fd Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:40.999981 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 06:27:41 crc kubenswrapper[4818]: W1203 06:27:41.011859 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-7320008d0d247a5fca654c5e67a6cfd1343d08775e5e5bcf04b1115652b41892 WatchSource:0}: Error finding container 7320008d0d247a5fca654c5e67a6cfd1343d08775e5e5bcf04b1115652b41892: Status 404 returned error can't find the container with id 7320008d0d247a5fca654c5e67a6cfd1343d08775e5e5bcf04b1115652b41892 Dec 03 06:27:41 crc kubenswrapper[4818]: W1203 06:27:41.014258 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-0c6c409278f3184f978b0bd27bfcc0fa20587d8de6284f6abc028ad800579d3d WatchSource:0}: Error finding container 0c6c409278f3184f978b0bd27bfcc0fa20587d8de6284f6abc028ad800579d3d: Status 404 returned error can't find the container with id 0c6c409278f3184f978b0bd27bfcc0fa20587d8de6284f6abc028ad800579d3d Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.260295 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:27:41 crc kubenswrapper[4818]: E1203 06:27:41.260434 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:27:42.26039712 +0000 UTC m=+19.952005872 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.361969 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.362034 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.362056 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.362076 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:27:41 crc kubenswrapper[4818]: E1203 06:27:41.362131 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:27:41 crc kubenswrapper[4818]: E1203 06:27:41.362187 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:27:41 crc kubenswrapper[4818]: E1203 06:27:41.362204 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:27:41 crc kubenswrapper[4818]: E1203 06:27:41.362215 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:41 crc kubenswrapper[4818]: E1203 06:27:41.362235 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:42.362216215 +0000 UTC m=+20.053825007 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:27:41 crc kubenswrapper[4818]: E1203 06:27:41.362264 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:42.362248065 +0000 UTC m=+20.053856857 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:41 crc kubenswrapper[4818]: E1203 06:27:41.362326 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:27:41 crc kubenswrapper[4818]: E1203 06:27:41.362351 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:42.362344045 +0000 UTC m=+20.053952877 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:27:41 crc kubenswrapper[4818]: E1203 06:27:41.362399 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:27:41 crc kubenswrapper[4818]: E1203 06:27:41.362408 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:27:41 crc kubenswrapper[4818]: E1203 06:27:41.362416 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:41 crc kubenswrapper[4818]: E1203 06:27:41.362439 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:42.362431694 +0000 UTC m=+20.054040556 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.736758 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.736860 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:27:41 crc kubenswrapper[4818]: E1203 06:27:41.736909 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:27:41 crc kubenswrapper[4818]: E1203 06:27:41.736981 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.835236 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"0c6c409278f3184f978b0bd27bfcc0fa20587d8de6284f6abc028ad800579d3d"} Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.837429 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc"} Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.837455 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac"} Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.837465 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7320008d0d247a5fca654c5e67a6cfd1343d08775e5e5bcf04b1115652b41892"} Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.839859 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b"} Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.839928 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d7a42366b6cb1d4363e8ed0d5b04fd55a1d4b54074fce899c5525714259439fd"} Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.845491 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.847913 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde"} Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.848248 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.873635 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.892694 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.906922 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.921494 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.933635 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.947149 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.965451 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.981482 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:41 crc kubenswrapper[4818]: I1203 06:27:41.996410 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:41Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.007315 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.018134 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.028572 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.040006 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.053893 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.268676 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:27:42 crc kubenswrapper[4818]: E1203 06:27:42.268952 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:27:44.2689202 +0000 UTC m=+21.960528952 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.369447 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.369507 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.369536 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.369555 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:42 crc kubenswrapper[4818]: E1203 06:27:42.369623 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:27:42 crc kubenswrapper[4818]: E1203 06:27:42.369692 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:27:42 crc kubenswrapper[4818]: E1203 06:27:42.369706 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:42 crc kubenswrapper[4818]: E1203 06:27:42.369740 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:27:42 crc kubenswrapper[4818]: E1203 06:27:42.369755 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:27:42 crc kubenswrapper[4818]: E1203 06:27:42.369803 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:27:42 crc kubenswrapper[4818]: E1203 06:27:42.369765 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:44.369745427 +0000 UTC m=+22.061354189 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:42 crc kubenswrapper[4818]: E1203 06:27:42.369853 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:42 crc kubenswrapper[4818]: E1203 06:27:42.369631 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:27:42 crc kubenswrapper[4818]: E1203 06:27:42.369881 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:44.369860317 +0000 UTC m=+22.061469119 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:27:42 crc kubenswrapper[4818]: E1203 06:27:42.369918 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:44.369894217 +0000 UTC m=+22.061503019 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:42 crc kubenswrapper[4818]: E1203 06:27:42.369945 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:44.369933997 +0000 UTC m=+22.061542849 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.740043 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:27:42 crc kubenswrapper[4818]: E1203 06:27:42.740841 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.747217 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.747741 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.748910 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.749500 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.750405 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.750929 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.751474 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.752419 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.753006 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.753862 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.754489 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.755723 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.756223 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.756704 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.757652 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.758147 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.759105 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.759455 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.760038 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.760993 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.761445 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.761656 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.762352 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.762760 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.763843 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.764309 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.764907 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.766009 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.766440 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.767440 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.768008 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.769054 4818 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.769173 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.770737 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.771751 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.772159 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.773697 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.774344 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.775197 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.775789 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.776766 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.777296 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.778196 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.778788 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.779648 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.779911 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.780394 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.781268 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.781728 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.783048 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.783693 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.784617 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.785090 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.786067 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.786606 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.787216 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.796835 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.809094 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.824097 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.844697 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:42 crc kubenswrapper[4818]: I1203 06:27:42.859133 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.531591 4818 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.534058 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.534102 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.534114 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.534174 4818 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.539956 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.543701 4818 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.544028 4818 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.545428 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.545464 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.545475 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.545490 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.545504 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:43Z","lastTransitionTime":"2025-12-03T06:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.554791 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.559036 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: E1203 06:27:43.562552 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.566985 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.567024 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.567036 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.567051 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.567063 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:43Z","lastTransitionTime":"2025-12-03T06:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.574695 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: E1203 06:27:43.579745 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.583376 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.583411 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.583421 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.583436 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.583446 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:43Z","lastTransitionTime":"2025-12-03T06:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.588174 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: E1203 06:27:43.598945 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.601577 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.602825 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.602863 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.602874 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.602890 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.602903 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:43Z","lastTransitionTime":"2025-12-03T06:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:43 crc kubenswrapper[4818]: E1203 06:27:43.615197 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.616779 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.618998 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.619038 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.619048 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.619063 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.619076 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:43Z","lastTransitionTime":"2025-12-03T06:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.628789 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: E1203 06:27:43.630718 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: E1203 06:27:43.630891 4818 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.632342 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.632385 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.632395 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.632412 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.632422 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:43Z","lastTransitionTime":"2025-12-03T06:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.632525 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.641425 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.657319 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.668190 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.678657 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.688221 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.705236 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.716192 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.726227 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.735214 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.735266 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.735279 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.735295 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.735307 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:43Z","lastTransitionTime":"2025-12-03T06:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.737408 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.737415 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:27:43 crc kubenswrapper[4818]: E1203 06:27:43.737534 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:27:43 crc kubenswrapper[4818]: E1203 06:27:43.737630 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.739850 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.837937 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.838009 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.838032 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.838062 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.838084 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:43Z","lastTransitionTime":"2025-12-03T06:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.853743 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14"} Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.865631 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.879266 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.892226 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.904926 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.917838 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.928338 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.940498 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.940535 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.940543 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.940558 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.940567 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:43Z","lastTransitionTime":"2025-12-03T06:27:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.945978 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:43 crc kubenswrapper[4818]: I1203 06:27:43.957968 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.043622 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.043695 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.043713 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.043738 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.043758 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:44Z","lastTransitionTime":"2025-12-03T06:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.146456 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.146498 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.146509 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.146522 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.146532 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:44Z","lastTransitionTime":"2025-12-03T06:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.248811 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.249011 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.249035 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.249059 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.249077 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:44Z","lastTransitionTime":"2025-12-03T06:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.287366 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:27:44 crc kubenswrapper[4818]: E1203 06:27:44.287501 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:27:48.287480625 +0000 UTC m=+25.979089377 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.351603 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.351639 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.351652 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.351667 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.351678 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:44Z","lastTransitionTime":"2025-12-03T06:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.388328 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.388392 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.388432 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.388487 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:44 crc kubenswrapper[4818]: E1203 06:27:44.388600 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:27:44 crc kubenswrapper[4818]: E1203 06:27:44.388653 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:27:44 crc kubenswrapper[4818]: E1203 06:27:44.388680 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:27:44 crc kubenswrapper[4818]: E1203 06:27:44.388722 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:27:44 crc kubenswrapper[4818]: E1203 06:27:44.388745 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:44 crc kubenswrapper[4818]: E1203 06:27:44.388745 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:27:44 crc kubenswrapper[4818]: E1203 06:27:44.388794 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:27:44 crc kubenswrapper[4818]: E1203 06:27:44.388884 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:44 crc kubenswrapper[4818]: E1203 06:27:44.388761 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:48.388727291 +0000 UTC m=+26.080336083 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:27:44 crc kubenswrapper[4818]: E1203 06:27:44.389016 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:48.388963321 +0000 UTC m=+26.080572113 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:27:44 crc kubenswrapper[4818]: E1203 06:27:44.389054 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:48.389038951 +0000 UTC m=+26.080647733 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:44 crc kubenswrapper[4818]: E1203 06:27:44.389077 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:48.3890654 +0000 UTC m=+26.080674192 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.453964 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.454010 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.454022 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.454038 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.454052 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:44Z","lastTransitionTime":"2025-12-03T06:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.556946 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.556983 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.556992 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.557006 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.557016 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:44Z","lastTransitionTime":"2025-12-03T06:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.659620 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.659655 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.659664 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.659677 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.659686 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:44Z","lastTransitionTime":"2025-12-03T06:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.736951 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:27:44 crc kubenswrapper[4818]: E1203 06:27:44.737130 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.761795 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.761900 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.761918 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.761943 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.761961 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:44Z","lastTransitionTime":"2025-12-03T06:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.864879 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.864932 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.864949 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.864970 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.864989 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:44Z","lastTransitionTime":"2025-12-03T06:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.968055 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.968155 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.968183 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.968211 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:44 crc kubenswrapper[4818]: I1203 06:27:44.968234 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:44Z","lastTransitionTime":"2025-12-03T06:27:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.071089 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.071141 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.071156 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.071176 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.071191 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:45Z","lastTransitionTime":"2025-12-03T06:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.173803 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.173885 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.173908 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.173933 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.173949 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:45Z","lastTransitionTime":"2025-12-03T06:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.279416 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.279465 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.279477 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.279492 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.279506 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:45Z","lastTransitionTime":"2025-12-03T06:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.369698 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.376552 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.381571 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.381640 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.381666 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.381697 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.381721 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:45Z","lastTransitionTime":"2025-12-03T06:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.381748 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.389004 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.405587 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.421793 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.444479 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.462258 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.476219 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.485360 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.485429 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.485438 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.485455 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.485467 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:45Z","lastTransitionTime":"2025-12-03T06:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.493116 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.516701 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.536297 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.550846 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.567450 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.580614 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.587727 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.587770 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.587784 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.587804 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.587852 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:45Z","lastTransitionTime":"2025-12-03T06:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.592163 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.603205 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.616482 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.631760 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.646384 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.690089 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.690129 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.690140 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.690344 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.690397 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:45Z","lastTransitionTime":"2025-12-03T06:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.736959 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.736987 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:27:45 crc kubenswrapper[4818]: E1203 06:27:45.737134 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:27:45 crc kubenswrapper[4818]: E1203 06:27:45.737198 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.792941 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.792976 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.792986 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.793001 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.793012 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:45Z","lastTransitionTime":"2025-12-03T06:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.895680 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.895715 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.895724 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.895737 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.895745 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:45Z","lastTransitionTime":"2025-12-03T06:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.997912 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.997962 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.997985 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.998004 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:45 crc kubenswrapper[4818]: I1203 06:27:45.998017 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:45Z","lastTransitionTime":"2025-12-03T06:27:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.100193 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.100230 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.100241 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.100257 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.100315 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:46Z","lastTransitionTime":"2025-12-03T06:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.202905 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.202949 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.202963 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.202981 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.202993 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:46Z","lastTransitionTime":"2025-12-03T06:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.255606 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-m4v69"] Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.255890 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-m4v69" Dec 03 06:27:46 crc kubenswrapper[4818]: W1203 06:27:46.257101 4818 reflector.go:561] object-"openshift-dns"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Dec 03 06:27:46 crc kubenswrapper[4818]: E1203 06:27:46.257143 4818 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.258047 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.258477 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.279985 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.296420 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.305103 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.305126 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.305135 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.305146 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.305154 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:46Z","lastTransitionTime":"2025-12-03T06:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.314515 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.327671 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.341102 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.351147 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.361203 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.377284 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.395982 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.406969 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.407005 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.407014 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.407029 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.407043 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:46Z","lastTransitionTime":"2025-12-03T06:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.408316 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1b5b453f-6556-4a63-ace6-494d845ff04b-hosts-file\") pod \"node-resolver-m4v69\" (UID: \"1b5b453f-6556-4a63-ace6-494d845ff04b\") " pod="openshift-dns/node-resolver-m4v69" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.408381 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pf6mj\" (UniqueName: \"kubernetes.io/projected/1b5b453f-6556-4a63-ace6-494d845ff04b-kube-api-access-pf6mj\") pod \"node-resolver-m4v69\" (UID: \"1b5b453f-6556-4a63-ace6-494d845ff04b\") " pod="openshift-dns/node-resolver-m4v69" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.409125 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.509113 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pf6mj\" (UniqueName: \"kubernetes.io/projected/1b5b453f-6556-4a63-ace6-494d845ff04b-kube-api-access-pf6mj\") pod \"node-resolver-m4v69\" (UID: \"1b5b453f-6556-4a63-ace6-494d845ff04b\") " pod="openshift-dns/node-resolver-m4v69" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.509184 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.509207 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.509215 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.509187 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1b5b453f-6556-4a63-ace6-494d845ff04b-hosts-file\") pod \"node-resolver-m4v69\" (UID: \"1b5b453f-6556-4a63-ace6-494d845ff04b\") " pod="openshift-dns/node-resolver-m4v69" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.509268 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1b5b453f-6556-4a63-ace6-494d845ff04b-hosts-file\") pod \"node-resolver-m4v69\" (UID: \"1b5b453f-6556-4a63-ace6-494d845ff04b\") " pod="openshift-dns/node-resolver-m4v69" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.509231 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.509352 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:46Z","lastTransitionTime":"2025-12-03T06:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.611861 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.611900 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.611911 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.611948 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.611959 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:46Z","lastTransitionTime":"2025-12-03T06:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.640157 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-ctpzs"] Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.640407 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-pjpw7"] Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.640421 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.640972 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.641163 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-t9fzk"] Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.641535 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.645063 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.645191 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.645369 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.645397 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.645638 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.645772 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.645910 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.646227 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.646431 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.646518 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.646628 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.648159 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.681665 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.713932 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.714186 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.714321 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.714415 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.714498 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:46Z","lastTransitionTime":"2025-12-03T06:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.725469 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.736884 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:27:46 crc kubenswrapper[4818]: E1203 06:27:46.737003 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.762106 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.811998 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812091 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/10740ef3-c8fb-4663-86ef-b46c1f969ffb-cni-binary-copy\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812129 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-var-lib-cni-multus\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812153 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/10740ef3-c8fb-4663-86ef-b46c1f969ffb-system-cni-dir\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812166 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-multus-cni-dir\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812182 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-run-multus-certs\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812209 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-system-cni-dir\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812226 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dd398f0d-786f-44e7-a64b-7cf7516d5d54-proxy-tls\") pod \"machine-config-daemon-t9fzk\" (UID: \"dd398f0d-786f-44e7-a64b-7cf7516d5d54\") " pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812240 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qxnz\" (UniqueName: \"kubernetes.io/projected/10740ef3-c8fb-4663-86ef-b46c1f969ffb-kube-api-access-4qxnz\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812256 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/10740ef3-c8fb-4663-86ef-b46c1f969ffb-cnibin\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812271 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/10740ef3-c8fb-4663-86ef-b46c1f969ffb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812287 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-var-lib-kubelet\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812337 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7b25b836-b639-4111-bcea-af7cc791ea32-cni-binary-copy\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812359 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-hostroot\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812378 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trn6s\" (UniqueName: \"kubernetes.io/projected/dd398f0d-786f-44e7-a64b-7cf7516d5d54-kube-api-access-trn6s\") pod \"machine-config-daemon-t9fzk\" (UID: \"dd398f0d-786f-44e7-a64b-7cf7516d5d54\") " pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812426 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-multus-conf-dir\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812448 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/dd398f0d-786f-44e7-a64b-7cf7516d5d54-rootfs\") pod \"machine-config-daemon-t9fzk\" (UID: \"dd398f0d-786f-44e7-a64b-7cf7516d5d54\") " pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812467 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-cnibin\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812484 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/10740ef3-c8fb-4663-86ef-b46c1f969ffb-os-release\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812500 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/10740ef3-c8fb-4663-86ef-b46c1f969ffb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812515 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dd398f0d-786f-44e7-a64b-7cf7516d5d54-mcd-auth-proxy-config\") pod \"machine-config-daemon-t9fzk\" (UID: \"dd398f0d-786f-44e7-a64b-7cf7516d5d54\") " pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812532 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-var-lib-cni-bin\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812544 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7b25b836-b639-4111-bcea-af7cc791ea32-multus-daemon-config\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812558 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vx5p\" (UniqueName: \"kubernetes.io/projected/7b25b836-b639-4111-bcea-af7cc791ea32-kube-api-access-6vx5p\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812580 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-os-release\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812597 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-multus-socket-dir-parent\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812612 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-run-k8s-cni-cncf-io\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812626 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-run-netns\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.812640 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-etc-kubernetes\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.816083 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.816100 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.816108 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.816119 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.816128 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:46Z","lastTransitionTime":"2025-12-03T06:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.840585 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.880667 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.903939 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.912942 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/10740ef3-c8fb-4663-86ef-b46c1f969ffb-os-release\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.912986 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/10740ef3-c8fb-4663-86ef-b46c1f969ffb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913007 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dd398f0d-786f-44e7-a64b-7cf7516d5d54-mcd-auth-proxy-config\") pod \"machine-config-daemon-t9fzk\" (UID: \"dd398f0d-786f-44e7-a64b-7cf7516d5d54\") " pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913024 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vx5p\" (UniqueName: \"kubernetes.io/projected/7b25b836-b639-4111-bcea-af7cc791ea32-kube-api-access-6vx5p\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913042 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-var-lib-cni-bin\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913057 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7b25b836-b639-4111-bcea-af7cc791ea32-multus-daemon-config\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913071 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-multus-socket-dir-parent\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913085 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-run-k8s-cni-cncf-io\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913093 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/10740ef3-c8fb-4663-86ef-b46c1f969ffb-os-release\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913168 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-run-netns\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913100 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-run-netns\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913216 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-etc-kubernetes\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913255 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-os-release\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913281 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/10740ef3-c8fb-4663-86ef-b46c1f969ffb-cni-binary-copy\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913300 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-var-lib-cni-multus\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913331 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-multus-cni-dir\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913352 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-run-multus-certs\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913374 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/10740ef3-c8fb-4663-86ef-b46c1f969ffb-system-cni-dir\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913403 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-system-cni-dir\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913427 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dd398f0d-786f-44e7-a64b-7cf7516d5d54-proxy-tls\") pod \"machine-config-daemon-t9fzk\" (UID: \"dd398f0d-786f-44e7-a64b-7cf7516d5d54\") " pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913459 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qxnz\" (UniqueName: \"kubernetes.io/projected/10740ef3-c8fb-4663-86ef-b46c1f969ffb-kube-api-access-4qxnz\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913482 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-var-lib-kubelet\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913507 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/10740ef3-c8fb-4663-86ef-b46c1f969ffb-cnibin\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913529 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/10740ef3-c8fb-4663-86ef-b46c1f969ffb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913549 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7b25b836-b639-4111-bcea-af7cc791ea32-cni-binary-copy\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913571 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-hostroot\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913593 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trn6s\" (UniqueName: \"kubernetes.io/projected/dd398f0d-786f-44e7-a64b-7cf7516d5d54-kube-api-access-trn6s\") pod \"machine-config-daemon-t9fzk\" (UID: \"dd398f0d-786f-44e7-a64b-7cf7516d5d54\") " pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913618 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-multus-conf-dir\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913638 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/dd398f0d-786f-44e7-a64b-7cf7516d5d54-rootfs\") pod \"machine-config-daemon-t9fzk\" (UID: \"dd398f0d-786f-44e7-a64b-7cf7516d5d54\") " pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913671 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-cnibin\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913748 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-cnibin\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913782 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-etc-kubernetes\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.913809 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-var-lib-cni-bin\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914102 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/10740ef3-c8fb-4663-86ef-b46c1f969ffb-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914164 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-os-release\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914492 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-system-cni-dir\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914518 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-var-lib-cni-multus\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914545 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/10740ef3-c8fb-4663-86ef-b46c1f969ffb-system-cni-dir\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914594 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-multus-socket-dir-parent\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914630 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-run-k8s-cni-cncf-io\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914665 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dd398f0d-786f-44e7-a64b-7cf7516d5d54-mcd-auth-proxy-config\") pod \"machine-config-daemon-t9fzk\" (UID: \"dd398f0d-786f-44e7-a64b-7cf7516d5d54\") " pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914688 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7b25b836-b639-4111-bcea-af7cc791ea32-multus-daemon-config\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914720 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-multus-conf-dir\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914743 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/10740ef3-c8fb-4663-86ef-b46c1f969ffb-cni-binary-copy\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914765 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-multus-cni-dir\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914752 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/10740ef3-c8fb-4663-86ef-b46c1f969ffb-cnibin\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914754 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-hostroot\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914833 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-var-lib-kubelet\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914850 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/dd398f0d-786f-44e7-a64b-7cf7516d5d54-rootfs\") pod \"machine-config-daemon-t9fzk\" (UID: \"dd398f0d-786f-44e7-a64b-7cf7516d5d54\") " pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914953 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7b25b836-b639-4111-bcea-af7cc791ea32-host-run-multus-certs\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.914998 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7b25b836-b639-4111-bcea-af7cc791ea32-cni-binary-copy\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.916077 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/10740ef3-c8fb-4663-86ef-b46c1f969ffb-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.921349 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dd398f0d-786f-44e7-a64b-7cf7516d5d54-proxy-tls\") pod \"machine-config-daemon-t9fzk\" (UID: \"dd398f0d-786f-44e7-a64b-7cf7516d5d54\") " pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.923975 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.924007 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.924015 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.924030 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.924040 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:46Z","lastTransitionTime":"2025-12-03T06:27:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.931140 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.940542 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trn6s\" (UniqueName: \"kubernetes.io/projected/dd398f0d-786f-44e7-a64b-7cf7516d5d54-kube-api-access-trn6s\") pod \"machine-config-daemon-t9fzk\" (UID: \"dd398f0d-786f-44e7-a64b-7cf7516d5d54\") " pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.942084 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qxnz\" (UniqueName: \"kubernetes.io/projected/10740ef3-c8fb-4663-86ef-b46c1f969ffb-kube-api-access-4qxnz\") pod \"multus-additional-cni-plugins-pjpw7\" (UID: \"10740ef3-c8fb-4663-86ef-b46c1f969ffb\") " pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.942272 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vx5p\" (UniqueName: \"kubernetes.io/projected/7b25b836-b639-4111-bcea-af7cc791ea32-kube-api-access-6vx5p\") pod \"multus-ctpzs\" (UID: \"7b25b836-b639-4111-bcea-af7cc791ea32\") " pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.953947 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.957151 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-ctpzs" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.965673 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" Dec 03 06:27:46 crc kubenswrapper[4818]: W1203 06:27:46.969594 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b25b836_b639_4111_bcea_af7cc791ea32.slice/crio-32171887884d0f572826514513f0599b4905b855801431f9dc10f37e6b54653d WatchSource:0}: Error finding container 32171887884d0f572826514513f0599b4905b855801431f9dc10f37e6b54653d: Status 404 returned error can't find the container with id 32171887884d0f572826514513f0599b4905b855801431f9dc10f37e6b54653d Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.970949 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.971216 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: W1203 06:27:46.982756 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10740ef3_c8fb_4663_86ef_b46c1f969ffb.slice/crio-c961e3c9f3116e025778c794c61aff0c61c7ee8784cb47b2528f70da779e063e WatchSource:0}: Error finding container c961e3c9f3116e025778c794c61aff0c61c7ee8784cb47b2528f70da779e063e: Status 404 returned error can't find the container with id c961e3c9f3116e025778c794c61aff0c61c7ee8784cb47b2528f70da779e063e Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.985267 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:46 crc kubenswrapper[4818]: W1203 06:27:46.985885 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd398f0d_786f_44e7_a64b_7cf7516d5d54.slice/crio-289e354b059a579008cf15197f97982bf3bd37af7113e2d6e37d62620a386204 WatchSource:0}: Error finding container 289e354b059a579008cf15197f97982bf3bd37af7113e2d6e37d62620a386204: Status 404 returned error can't find the container with id 289e354b059a579008cf15197f97982bf3bd37af7113e2d6e37d62620a386204 Dec 03 06:27:46 crc kubenswrapper[4818]: I1203 06:27:46.997885 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:46Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.010304 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.012446 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ktm8k"] Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.013234 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.015292 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.015349 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.015467 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.015480 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.015302 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.015669 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.015848 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.023666 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.033430 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.033448 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.033456 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.033469 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.033478 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:47Z","lastTransitionTime":"2025-12-03T06:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.044553 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.058110 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.070885 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.081915 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.093717 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.110061 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115041 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-slash\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115068 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-openvswitch\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115087 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-systemd\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115103 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-log-socket\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115171 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-ovnkube-config\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115251 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrtqq\" (UniqueName: \"kubernetes.io/projected/a1e30fe6-0b94-4c95-b941-c0623729e123-kube-api-access-hrtqq\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115298 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-env-overrides\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115329 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-run-ovn-kubernetes\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115350 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-ovnkube-script-lib\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115378 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-etc-openvswitch\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115397 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-cni-netd\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115428 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-ovn\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115454 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-kubelet\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115531 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a1e30fe6-0b94-4c95-b941-c0623729e123-ovn-node-metrics-cert\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115570 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-var-lib-openvswitch\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115610 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-run-netns\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115628 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-node-log\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115727 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-systemd-units\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115756 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.115852 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-cni-bin\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.123270 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.136017 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.136068 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.136105 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.136117 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.136137 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.136149 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:47Z","lastTransitionTime":"2025-12-03T06:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.152017 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.168870 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.184145 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.197335 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.215049 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216535 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-run-netns\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216573 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-node-log\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216616 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-systemd-units\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216638 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216648 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-run-netns\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216660 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-cni-bin\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216704 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-slash\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216721 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-openvswitch\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216726 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-node-log\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216757 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-systemd\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216704 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-cni-bin\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216792 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-log-socket\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216771 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-systemd-units\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216803 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216829 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-openvswitch\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216775 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-log-socket\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216843 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-slash\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216858 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-ovnkube-config\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216862 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-systemd\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216904 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrtqq\" (UniqueName: \"kubernetes.io/projected/a1e30fe6-0b94-4c95-b941-c0623729e123-kube-api-access-hrtqq\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216922 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-env-overrides\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216942 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-run-ovn-kubernetes\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216960 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-ovnkube-script-lib\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.216978 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-etc-openvswitch\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.217013 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-cni-netd\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.217030 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-ovn\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.217046 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-kubelet\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.217069 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a1e30fe6-0b94-4c95-b941-c0623729e123-ovn-node-metrics-cert\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.217084 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-var-lib-openvswitch\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.217130 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-var-lib-openvswitch\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.217606 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-ovnkube-config\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.217649 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-cni-netd\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.217676 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-run-ovn-kubernetes\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.217747 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-env-overrides\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.217802 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-ovn\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.217850 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-kubelet\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.218212 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-ovnkube-script-lib\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.218261 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-etc-openvswitch\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.224504 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a1e30fe6-0b94-4c95-b941-c0623729e123-ovn-node-metrics-cert\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.234933 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrtqq\" (UniqueName: \"kubernetes.io/projected/a1e30fe6-0b94-4c95-b941-c0623729e123-kube-api-access-hrtqq\") pod \"ovnkube-node-ktm8k\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.237808 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.238308 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.238346 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.238356 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.238371 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.238383 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:47Z","lastTransitionTime":"2025-12-03T06:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.253398 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.263468 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.282057 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.294917 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.305251 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.314546 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.325779 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.335012 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.336855 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.340620 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.340649 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.340659 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.340674 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.340683 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:47Z","lastTransitionTime":"2025-12-03T06:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:47 crc kubenswrapper[4818]: W1203 06:27:47.347171 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1e30fe6_0b94_4c95_b941_c0623729e123.slice/crio-cc3f0301bc894b4e2819c9d25e37cdeca35b35f3e7b4c11fb332c9c8653378e9 WatchSource:0}: Error finding container cc3f0301bc894b4e2819c9d25e37cdeca35b35f3e7b4c11fb332c9c8653378e9: Status 404 returned error can't find the container with id cc3f0301bc894b4e2819c9d25e37cdeca35b35f3e7b4c11fb332c9c8653378e9 Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.351459 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.366147 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.368772 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.376507 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pf6mj\" (UniqueName: \"kubernetes.io/projected/1b5b453f-6556-4a63-ace6-494d845ff04b-kube-api-access-pf6mj\") pod \"node-resolver-m4v69\" (UID: \"1b5b453f-6556-4a63-ace6-494d845ff04b\") " pod="openshift-dns/node-resolver-m4v69" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.443361 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.443406 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.443417 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.443430 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.443439 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:47Z","lastTransitionTime":"2025-12-03T06:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.467939 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-m4v69" Dec 03 06:27:47 crc kubenswrapper[4818]: W1203 06:27:47.480733 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b5b453f_6556_4a63_ace6_494d845ff04b.slice/crio-d94f84d67f1c48561d676df81ebc6c0fdba661351db8d8cef20ea2399a68fef2 WatchSource:0}: Error finding container d94f84d67f1c48561d676df81ebc6c0fdba661351db8d8cef20ea2399a68fef2: Status 404 returned error can't find the container with id d94f84d67f1c48561d676df81ebc6c0fdba661351db8d8cef20ea2399a68fef2 Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.546958 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.547012 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.547027 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.547046 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.547059 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:47Z","lastTransitionTime":"2025-12-03T06:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.651960 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.652365 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.652379 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.652399 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.652411 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:47Z","lastTransitionTime":"2025-12-03T06:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.737002 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:47 crc kubenswrapper[4818]: E1203 06:27:47.737134 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.737002 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:27:47 crc kubenswrapper[4818]: E1203 06:27:47.737404 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.754990 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.755029 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.755208 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.755223 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.755234 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:47Z","lastTransitionTime":"2025-12-03T06:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.857928 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.857991 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.858004 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.858027 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.858043 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:47Z","lastTransitionTime":"2025-12-03T06:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.863590 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-m4v69" event={"ID":"1b5b453f-6556-4a63-ace6-494d845ff04b","Type":"ContainerStarted","Data":"d94f84d67f1c48561d676df81ebc6c0fdba661351db8d8cef20ea2399a68fef2"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.865275 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ctpzs" event={"ID":"7b25b836-b639-4111-bcea-af7cc791ea32","Type":"ContainerStarted","Data":"3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.865321 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ctpzs" event={"ID":"7b25b836-b639-4111-bcea-af7cc791ea32","Type":"ContainerStarted","Data":"32171887884d0f572826514513f0599b4905b855801431f9dc10f37e6b54653d"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.867082 4818 generic.go:334] "Generic (PLEG): container finished" podID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerID="16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16" exitCode=0 Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.867123 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerDied","Data":"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.867187 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerStarted","Data":"cc3f0301bc894b4e2819c9d25e37cdeca35b35f3e7b4c11fb332c9c8653378e9"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.869183 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.869215 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.869226 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"289e354b059a579008cf15197f97982bf3bd37af7113e2d6e37d62620a386204"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.870560 4818 generic.go:334] "Generic (PLEG): container finished" podID="10740ef3-c8fb-4663-86ef-b46c1f969ffb" containerID="c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108" exitCode=0 Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.870601 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" event={"ID":"10740ef3-c8fb-4663-86ef-b46c1f969ffb","Type":"ContainerDied","Data":"c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.870634 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" event={"ID":"10740ef3-c8fb-4663-86ef-b46c1f969ffb","Type":"ContainerStarted","Data":"c961e3c9f3116e025778c794c61aff0c61c7ee8784cb47b2528f70da779e063e"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.881780 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.894146 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.906915 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.928192 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.940151 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.952673 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.960779 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.961433 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.961445 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.961460 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.961471 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:47Z","lastTransitionTime":"2025-12-03T06:27:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.969764 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.984059 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:47 crc kubenswrapper[4818]: I1203 06:27:47.994654 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:47Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.016142 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.030151 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.042868 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.055422 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.064265 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.064311 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.064322 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.064336 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.064348 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:48Z","lastTransitionTime":"2025-12-03T06:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.069101 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.082224 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.094774 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.115870 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.132665 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.147890 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.162340 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.166050 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.166190 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.166245 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.166301 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.166352 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:48Z","lastTransitionTime":"2025-12-03T06:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.175612 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.216308 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.230779 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.254003 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.268249 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.268601 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.268657 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.268668 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.268685 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.268697 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:48Z","lastTransitionTime":"2025-12-03T06:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.284331 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.299020 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.312485 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.328857 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:27:48 crc kubenswrapper[4818]: E1203 06:27:48.329010 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:27:56.328986341 +0000 UTC m=+34.020595093 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.371887 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.371925 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.371935 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.371947 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.371958 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:48Z","lastTransitionTime":"2025-12-03T06:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.430580 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.430650 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.430682 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:27:48 crc kubenswrapper[4818]: E1203 06:27:48.430704 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:27:48 crc kubenswrapper[4818]: E1203 06:27:48.430759 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:56.430745198 +0000 UTC m=+34.122353950 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:27:48 crc kubenswrapper[4818]: E1203 06:27:48.430800 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:27:48 crc kubenswrapper[4818]: E1203 06:27:48.430839 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:27:48 crc kubenswrapper[4818]: E1203 06:27:48.430861 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:48 crc kubenswrapper[4818]: E1203 06:27:48.430896 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:56.430884041 +0000 UTC m=+34.122492803 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:48 crc kubenswrapper[4818]: E1203 06:27:48.430954 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:27:48 crc kubenswrapper[4818]: E1203 06:27:48.430983 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:56.430974724 +0000 UTC m=+34.122583486 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:27:48 crc kubenswrapper[4818]: E1203 06:27:48.431035 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.430708 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:27:48 crc kubenswrapper[4818]: E1203 06:27:48.431047 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:27:48 crc kubenswrapper[4818]: E1203 06:27:48.431056 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:48 crc kubenswrapper[4818]: E1203 06:27:48.431081 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:27:56.431073496 +0000 UTC m=+34.122682258 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.473930 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.473973 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.473982 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.473997 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.474008 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:48Z","lastTransitionTime":"2025-12-03T06:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.576025 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.576061 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.576070 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.576084 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.576095 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:48Z","lastTransitionTime":"2025-12-03T06:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.678669 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.678721 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.678733 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.678755 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.678769 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:48Z","lastTransitionTime":"2025-12-03T06:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.737569 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:27:48 crc kubenswrapper[4818]: E1203 06:27:48.737833 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.781945 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.782023 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.782051 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.782081 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.782098 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:48Z","lastTransitionTime":"2025-12-03T06:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.874464 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-m4v69" event={"ID":"1b5b453f-6556-4a63-ace6-494d845ff04b","Type":"ContainerStarted","Data":"e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653"} Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.877439 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerStarted","Data":"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc"} Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.877474 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerStarted","Data":"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643"} Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.877486 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerStarted","Data":"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a"} Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.877496 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerStarted","Data":"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0"} Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.877507 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerStarted","Data":"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61"} Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.877519 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerStarted","Data":"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0"} Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.879089 4818 generic.go:334] "Generic (PLEG): container finished" podID="10740ef3-c8fb-4663-86ef-b46c1f969ffb" containerID="dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8" exitCode=0 Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.879120 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" event={"ID":"10740ef3-c8fb-4663-86ef-b46c1f969ffb","Type":"ContainerDied","Data":"dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8"} Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.884368 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.884399 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.884407 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.884420 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.884432 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:48Z","lastTransitionTime":"2025-12-03T06:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.895241 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.907571 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.929901 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.944195 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.955762 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.967990 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.981647 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.987761 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.987805 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.987830 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.987848 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:48 crc kubenswrapper[4818]: I1203 06:27:48.987860 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:48Z","lastTransitionTime":"2025-12-03T06:27:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.020328 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.048755 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.067425 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.082714 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.090252 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.090280 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.090288 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.090300 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.090309 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:49Z","lastTransitionTime":"2025-12-03T06:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.095708 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.108675 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.127645 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.138691 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.152866 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.164617 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.182713 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.192288 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.192324 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.192333 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.192346 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.192357 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:49Z","lastTransitionTime":"2025-12-03T06:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.195998 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.213089 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.226982 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.243025 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.258459 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.272294 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.293758 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.294932 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.294977 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.294993 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.295013 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.295026 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:49Z","lastTransitionTime":"2025-12-03T06:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.309702 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.324995 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.336711 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.397984 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.398020 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.398030 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.398046 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.398058 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:49Z","lastTransitionTime":"2025-12-03T06:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.501534 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.501590 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.501604 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.501625 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.501638 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:49Z","lastTransitionTime":"2025-12-03T06:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.603690 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.603717 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.603726 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.603738 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.603747 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:49Z","lastTransitionTime":"2025-12-03T06:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.705549 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.705595 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.705610 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.705622 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.705637 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:49Z","lastTransitionTime":"2025-12-03T06:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.737133 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:49 crc kubenswrapper[4818]: E1203 06:27:49.737316 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.737330 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:27:49 crc kubenswrapper[4818]: E1203 06:27:49.737535 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.808564 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.808621 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.808635 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.808651 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.808663 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:49Z","lastTransitionTime":"2025-12-03T06:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.884044 4818 generic.go:334] "Generic (PLEG): container finished" podID="10740ef3-c8fb-4663-86ef-b46c1f969ffb" containerID="afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d" exitCode=0 Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.884246 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" event={"ID":"10740ef3-c8fb-4663-86ef-b46c1f969ffb","Type":"ContainerDied","Data":"afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d"} Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.897421 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.910694 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.910737 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.910746 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.910762 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.910772 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:49Z","lastTransitionTime":"2025-12-03T06:27:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.912164 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.926491 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.940532 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.957975 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.972450 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.972773 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-2ds8v"] Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.973390 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2ds8v" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.975529 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.976039 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.976264 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.977029 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 06:27:49 crc kubenswrapper[4818]: I1203 06:27:49.991313 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.007065 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.013670 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.013729 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.013740 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.013759 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.013769 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:50Z","lastTransitionTime":"2025-12-03T06:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.020664 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.039006 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.056377 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.070557 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.080796 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.093559 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.110159 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.116771 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.116844 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.116860 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.116877 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.116890 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:50Z","lastTransitionTime":"2025-12-03T06:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.123901 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.141660 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.147445 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/84d9588b-f409-4fc0-9f8e-3ec975189e5c-serviceca\") pod \"node-ca-2ds8v\" (UID: \"84d9588b-f409-4fc0-9f8e-3ec975189e5c\") " pod="openshift-image-registry/node-ca-2ds8v" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.147496 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84d9588b-f409-4fc0-9f8e-3ec975189e5c-host\") pod \"node-ca-2ds8v\" (UID: \"84d9588b-f409-4fc0-9f8e-3ec975189e5c\") " pod="openshift-image-registry/node-ca-2ds8v" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.147539 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq84n\" (UniqueName: \"kubernetes.io/projected/84d9588b-f409-4fc0-9f8e-3ec975189e5c-kube-api-access-xq84n\") pod \"node-ca-2ds8v\" (UID: \"84d9588b-f409-4fc0-9f8e-3ec975189e5c\") " pod="openshift-image-registry/node-ca-2ds8v" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.164527 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.180062 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.193041 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.204538 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.214242 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.218741 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.218803 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.218853 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.218878 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.218897 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:50Z","lastTransitionTime":"2025-12-03T06:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.228098 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.241987 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.247960 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq84n\" (UniqueName: \"kubernetes.io/projected/84d9588b-f409-4fc0-9f8e-3ec975189e5c-kube-api-access-xq84n\") pod \"node-ca-2ds8v\" (UID: \"84d9588b-f409-4fc0-9f8e-3ec975189e5c\") " pod="openshift-image-registry/node-ca-2ds8v" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.248030 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/84d9588b-f409-4fc0-9f8e-3ec975189e5c-serviceca\") pod \"node-ca-2ds8v\" (UID: \"84d9588b-f409-4fc0-9f8e-3ec975189e5c\") " pod="openshift-image-registry/node-ca-2ds8v" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.248056 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84d9588b-f409-4fc0-9f8e-3ec975189e5c-host\") pod \"node-ca-2ds8v\" (UID: \"84d9588b-f409-4fc0-9f8e-3ec975189e5c\") " pod="openshift-image-registry/node-ca-2ds8v" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.248120 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84d9588b-f409-4fc0-9f8e-3ec975189e5c-host\") pod \"node-ca-2ds8v\" (UID: \"84d9588b-f409-4fc0-9f8e-3ec975189e5c\") " pod="openshift-image-registry/node-ca-2ds8v" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.249461 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/84d9588b-f409-4fc0-9f8e-3ec975189e5c-serviceca\") pod \"node-ca-2ds8v\" (UID: \"84d9588b-f409-4fc0-9f8e-3ec975189e5c\") " pod="openshift-image-registry/node-ca-2ds8v" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.254627 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.274776 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq84n\" (UniqueName: \"kubernetes.io/projected/84d9588b-f409-4fc0-9f8e-3ec975189e5c-kube-api-access-xq84n\") pod \"node-ca-2ds8v\" (UID: \"84d9588b-f409-4fc0-9f8e-3ec975189e5c\") " pod="openshift-image-registry/node-ca-2ds8v" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.276429 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.288174 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2ds8v" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.290376 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: W1203 06:27:50.300680 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84d9588b_f409_4fc0_9f8e_3ec975189e5c.slice/crio-681fba4e7bb732e76b83ba82ccfa16d3788757b5120130cf806a461a645ceca7 WatchSource:0}: Error finding container 681fba4e7bb732e76b83ba82ccfa16d3788757b5120130cf806a461a645ceca7: Status 404 returned error can't find the container with id 681fba4e7bb732e76b83ba82ccfa16d3788757b5120130cf806a461a645ceca7 Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.303090 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.313407 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.320734 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.320761 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.320769 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.320781 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.320791 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:50Z","lastTransitionTime":"2025-12-03T06:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.423016 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.423048 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.423056 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.423069 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.423078 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:50Z","lastTransitionTime":"2025-12-03T06:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.525708 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.525747 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.525758 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.525773 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.525784 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:50Z","lastTransitionTime":"2025-12-03T06:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.634358 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.634499 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.634520 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.634549 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.634568 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:50Z","lastTransitionTime":"2025-12-03T06:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.736848 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.737303 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.737629 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.737641 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:50 crc kubenswrapper[4818]: E1203 06:27:50.737638 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.737660 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.737701 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:50Z","lastTransitionTime":"2025-12-03T06:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.839730 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.839779 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.839791 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.839806 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.839831 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:50Z","lastTransitionTime":"2025-12-03T06:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.891284 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerStarted","Data":"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7"} Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.894133 4818 generic.go:334] "Generic (PLEG): container finished" podID="10740ef3-c8fb-4663-86ef-b46c1f969ffb" containerID="e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189" exitCode=0 Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.894200 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" event={"ID":"10740ef3-c8fb-4663-86ef-b46c1f969ffb","Type":"ContainerDied","Data":"e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189"} Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.896633 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2ds8v" event={"ID":"84d9588b-f409-4fc0-9f8e-3ec975189e5c","Type":"ContainerStarted","Data":"981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32"} Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.896685 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2ds8v" event={"ID":"84d9588b-f409-4fc0-9f8e-3ec975189e5c","Type":"ContainerStarted","Data":"681fba4e7bb732e76b83ba82ccfa16d3788757b5120130cf806a461a645ceca7"} Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.924314 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.942676 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.942925 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.943042 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.943134 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.943225 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:50Z","lastTransitionTime":"2025-12-03T06:27:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.943863 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.966011 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.977382 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:50 crc kubenswrapper[4818]: I1203 06:27:50.993466 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.005755 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.018198 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.033612 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.046761 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.046840 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.046855 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.046874 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.046887 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:51Z","lastTransitionTime":"2025-12-03T06:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.047142 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.063874 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.077439 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.091669 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.131805 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.149309 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.149359 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.149374 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.149394 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.149408 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:51Z","lastTransitionTime":"2025-12-03T06:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.169052 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.181495 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.200064 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.214994 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.228748 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.243867 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.252485 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.252523 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.252534 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.252549 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.252560 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:51Z","lastTransitionTime":"2025-12-03T06:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.257839 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.271271 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.285605 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.297860 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.313007 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.325562 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.339404 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.353240 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.354827 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.354883 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.354897 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.354916 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.354929 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:51Z","lastTransitionTime":"2025-12-03T06:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.371649 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.410042 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.445115 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.457741 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.457780 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.457789 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.457806 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.457828 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:51Z","lastTransitionTime":"2025-12-03T06:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.561017 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.561058 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.561068 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.561084 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.561096 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:51Z","lastTransitionTime":"2025-12-03T06:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.663490 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.663759 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.663862 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.663961 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.664045 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:51Z","lastTransitionTime":"2025-12-03T06:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.737147 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:51 crc kubenswrapper[4818]: E1203 06:27:51.737793 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.737173 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:27:51 crc kubenswrapper[4818]: E1203 06:27:51.738091 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.766010 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.766228 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.766423 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.766621 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.766852 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:51Z","lastTransitionTime":"2025-12-03T06:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.870757 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.870810 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.870862 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.870891 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.870927 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:51Z","lastTransitionTime":"2025-12-03T06:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.904614 4818 generic.go:334] "Generic (PLEG): container finished" podID="10740ef3-c8fb-4663-86ef-b46c1f969ffb" containerID="7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc" exitCode=0 Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.904676 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" event={"ID":"10740ef3-c8fb-4663-86ef-b46c1f969ffb","Type":"ContainerDied","Data":"7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc"} Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.918342 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.934811 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.950434 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.973780 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.974049 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.974127 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.974210 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.974302 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:51Z","lastTransitionTime":"2025-12-03T06:27:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.976749 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:51 crc kubenswrapper[4818]: I1203 06:27:51.991900 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:51Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.006283 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.026871 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.047621 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.065348 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.076011 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.076430 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.076460 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.076470 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.076570 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.076589 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:52Z","lastTransitionTime":"2025-12-03T06:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.089526 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.103198 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.116283 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.129722 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.141716 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.179171 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.179222 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.179234 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.179253 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.179266 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:52Z","lastTransitionTime":"2025-12-03T06:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.284844 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.284878 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.284891 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.284905 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.284915 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:52Z","lastTransitionTime":"2025-12-03T06:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.387896 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.387937 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.387945 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.387959 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.387969 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:52Z","lastTransitionTime":"2025-12-03T06:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.491007 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.491046 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.491057 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.491071 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.491081 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:52Z","lastTransitionTime":"2025-12-03T06:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.594262 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.594310 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.594321 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.594335 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.594345 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:52Z","lastTransitionTime":"2025-12-03T06:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.697188 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.697223 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.697235 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.697251 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.697262 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:52Z","lastTransitionTime":"2025-12-03T06:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.737380 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:27:52 crc kubenswrapper[4818]: E1203 06:27:52.737563 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.752950 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.768509 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.790984 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.800169 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.800233 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.800254 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.800280 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.800303 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:52Z","lastTransitionTime":"2025-12-03T06:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.821314 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.836743 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.850642 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.871280 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.884387 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.895327 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.902761 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.902805 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.902836 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.902855 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.902867 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:52Z","lastTransitionTime":"2025-12-03T06:27:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.908708 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.911368 4818 generic.go:334] "Generic (PLEG): container finished" podID="10740ef3-c8fb-4663-86ef-b46c1f969ffb" containerID="d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f" exitCode=0 Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.911420 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" event={"ID":"10740ef3-c8fb-4663-86ef-b46c1f969ffb","Type":"ContainerDied","Data":"d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f"} Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.926467 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.944429 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.956046 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.969049 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.981981 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:52 crc kubenswrapper[4818]: I1203 06:27:52.996581 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.009557 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.010246 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.010268 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.010289 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.010306 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:53Z","lastTransitionTime":"2025-12-03T06:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.012019 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:53Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.027416 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:53Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.045998 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:53Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.067190 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:53Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.082252 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:53Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.106783 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:53Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.112904 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.112946 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.112957 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.112974 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.112985 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:53Z","lastTransitionTime":"2025-12-03T06:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.124294 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:53Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.138899 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:53Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.152665 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:53Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.165885 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:53Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.181374 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:53Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.195633 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:53Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.208772 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:53Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.215960 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.216001 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.216012 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.216030 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.216044 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:53Z","lastTransitionTime":"2025-12-03T06:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.245932 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:53Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.318487 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.318515 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.318529 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.318543 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.318552 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:53Z","lastTransitionTime":"2025-12-03T06:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.421027 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.421101 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.421121 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.421143 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.421208 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:53Z","lastTransitionTime":"2025-12-03T06:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.524189 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.524502 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.524514 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.524533 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.524551 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:53Z","lastTransitionTime":"2025-12-03T06:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.627141 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.627221 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.627245 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.627276 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.627298 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:53Z","lastTransitionTime":"2025-12-03T06:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.730189 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.730258 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.730277 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.730300 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.730320 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:53Z","lastTransitionTime":"2025-12-03T06:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.737517 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.737527 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:53 crc kubenswrapper[4818]: E1203 06:27:53.737669 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:27:53 crc kubenswrapper[4818]: E1203 06:27:53.737880 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.832533 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.832585 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.832597 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.832616 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.832628 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:53Z","lastTransitionTime":"2025-12-03T06:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.920917 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerStarted","Data":"20eb93f400d5792232c343625af736d163a5f32f5a469e25feaf0af6de1446c9"} Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.921442 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.926311 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" event={"ID":"10740ef3-c8fb-4663-86ef-b46c1f969ffb","Type":"ContainerStarted","Data":"0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16"} Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.935869 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.935895 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.935905 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.935920 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.935932 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:53Z","lastTransitionTime":"2025-12-03T06:27:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.943474 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:53Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.949605 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.960365 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:53Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:53 crc kubenswrapper[4818]: I1203 06:27:53.973954 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:53Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.006472 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb93f400d5792232c343625af736d163a5f32f5a469e25feaf0af6de1446c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.024304 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.026174 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.026227 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.026249 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.026279 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.026298 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:54Z","lastTransitionTime":"2025-12-03T06:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:54 crc kubenswrapper[4818]: E1203 06:27:54.040085 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.041770 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.044475 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.044531 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.044548 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.044575 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.044592 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:54Z","lastTransitionTime":"2025-12-03T06:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.054737 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: E1203 06:27:54.060500 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.068956 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.069001 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.069019 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.069042 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.069059 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:54Z","lastTransitionTime":"2025-12-03T06:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.075270 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: E1203 06:27:54.082638 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.085911 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.085975 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.085994 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.086021 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.086039 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:54Z","lastTransitionTime":"2025-12-03T06:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.089653 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: E1203 06:27:54.101357 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.104860 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.105110 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.105158 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.105171 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.105190 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.105203 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:54Z","lastTransitionTime":"2025-12-03T06:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:54 crc kubenswrapper[4818]: E1203 06:27:54.120607 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: E1203 06:27:54.120713 4818 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.122503 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.122529 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.122540 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.122557 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.122569 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:54Z","lastTransitionTime":"2025-12-03T06:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.133440 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.150518 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.168759 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.185932 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.196594 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.220692 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.225708 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.225767 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.225780 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.225799 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.225830 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:54Z","lastTransitionTime":"2025-12-03T06:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.239993 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.255563 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.268388 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.286679 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.303014 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.315917 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.329256 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.329324 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.329342 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.329368 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.329386 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:54Z","lastTransitionTime":"2025-12-03T06:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.332544 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.344461 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.363026 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.376468 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.394125 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.416296 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb93f400d5792232c343625af736d163a5f32f5a469e25feaf0af6de1446c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.432374 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.432408 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.432419 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.432435 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.432447 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:54Z","lastTransitionTime":"2025-12-03T06:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.433443 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.447799 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.535225 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.535269 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.535280 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.535297 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.535308 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:54Z","lastTransitionTime":"2025-12-03T06:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.637424 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.637454 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.637462 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.637474 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.637484 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:54Z","lastTransitionTime":"2025-12-03T06:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.738113 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:27:54 crc kubenswrapper[4818]: E1203 06:27:54.738383 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.739935 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.739974 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.739985 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.740001 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.740013 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:54Z","lastTransitionTime":"2025-12-03T06:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.842552 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.842642 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.842653 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.842672 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.842684 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:54Z","lastTransitionTime":"2025-12-03T06:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.929686 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.930226 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.945354 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.945401 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.945412 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.945430 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.945446 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:54Z","lastTransitionTime":"2025-12-03T06:27:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.951757 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.972082 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.984285 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:54 crc kubenswrapper[4818]: I1203 06:27:54.996463 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:54Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.012204 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.034530 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.047722 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.047772 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.047786 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.047809 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.047854 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:55Z","lastTransitionTime":"2025-12-03T06:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.052177 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.069793 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.090347 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.106789 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.128452 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.141557 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.149831 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.149870 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.149881 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.149898 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.149910 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:55Z","lastTransitionTime":"2025-12-03T06:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.153805 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.173028 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb93f400d5792232c343625af736d163a5f32f5a469e25feaf0af6de1446c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.186480 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.204272 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.252777 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.252822 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.252834 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.252849 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.252858 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:55Z","lastTransitionTime":"2025-12-03T06:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.355105 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.355161 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.355169 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.355182 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.355191 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:55Z","lastTransitionTime":"2025-12-03T06:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.457834 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.457873 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.457883 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.457900 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.457911 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:55Z","lastTransitionTime":"2025-12-03T06:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.559994 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.560032 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.560063 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.560081 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.560092 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:55Z","lastTransitionTime":"2025-12-03T06:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.663179 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.663234 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.663248 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.663270 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.663285 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:55Z","lastTransitionTime":"2025-12-03T06:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.737458 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.737479 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:27:55 crc kubenswrapper[4818]: E1203 06:27:55.737592 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:27:55 crc kubenswrapper[4818]: E1203 06:27:55.738265 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.766312 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.766363 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.766375 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.766396 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.766408 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:55Z","lastTransitionTime":"2025-12-03T06:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.870064 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.870130 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.870165 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.870204 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.870228 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:55Z","lastTransitionTime":"2025-12-03T06:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.933836 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovnkube-controller/0.log" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.937267 4818 generic.go:334] "Generic (PLEG): container finished" podID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerID="20eb93f400d5792232c343625af736d163a5f32f5a469e25feaf0af6de1446c9" exitCode=1 Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.937302 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerDied","Data":"20eb93f400d5792232c343625af736d163a5f32f5a469e25feaf0af6de1446c9"} Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.938156 4818 scope.go:117] "RemoveContainer" containerID="20eb93f400d5792232c343625af736d163a5f32f5a469e25feaf0af6de1446c9" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.955453 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.972727 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.973156 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.973170 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.973187 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.973073 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.973197 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:55Z","lastTransitionTime":"2025-12-03T06:27:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:55 crc kubenswrapper[4818]: I1203 06:27:55.993382 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:55Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.007626 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.022725 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.054429 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.068459 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.075565 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.075615 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.075633 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.075654 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.075667 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:56Z","lastTransitionTime":"2025-12-03T06:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.089976 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb93f400d5792232c343625af736d163a5f32f5a469e25feaf0af6de1446c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20eb93f400d5792232c343625af736d163a5f32f5a469e25feaf0af6de1446c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:27:55Z\\\",\\\"message\\\":\\\"or *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 06:27:55.468625 6107 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 06:27:55.468728 6107 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 06:27:55.469892 6107 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:27:55.469915 6107 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 06:27:55.469919 6107 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 06:27:55.469931 6107 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:27:55.469935 6107 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:27:55.469948 6107 factory.go:656] Stopping watch factory\\\\nI1203 06:27:55.469959 6107 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:27:55.469983 6107 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:27:55.469992 6107 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:27:55.469998 6107 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 06:27:55.470004 6107 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.100909 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.113239 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.126846 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.139732 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.151854 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.165599 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.178527 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.178574 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.178586 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.178605 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.178618 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:56Z","lastTransitionTime":"2025-12-03T06:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.188892 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.281668 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.281711 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.281720 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.281734 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.281744 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:56Z","lastTransitionTime":"2025-12-03T06:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.339085 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:27:56 crc kubenswrapper[4818]: E1203 06:27:56.339270 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:28:12.339251949 +0000 UTC m=+50.030860701 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.385025 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.385084 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.385102 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.385126 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.385142 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:56Z","lastTransitionTime":"2025-12-03T06:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.440007 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.440065 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.440098 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.440124 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:27:56 crc kubenswrapper[4818]: E1203 06:27:56.440241 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:27:56 crc kubenswrapper[4818]: E1203 06:27:56.440259 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:27:56 crc kubenswrapper[4818]: E1203 06:27:56.440273 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:56 crc kubenswrapper[4818]: E1203 06:27:56.440319 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:28:12.440303788 +0000 UTC m=+50.131912550 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:56 crc kubenswrapper[4818]: E1203 06:27:56.440610 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:27:56 crc kubenswrapper[4818]: E1203 06:27:56.440642 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:28:12.440632957 +0000 UTC m=+50.132241719 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:27:56 crc kubenswrapper[4818]: E1203 06:27:56.440694 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:27:56 crc kubenswrapper[4818]: E1203 06:27:56.440752 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:28:12.440726399 +0000 UTC m=+50.132335161 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:27:56 crc kubenswrapper[4818]: E1203 06:27:56.440780 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:27:56 crc kubenswrapper[4818]: E1203 06:27:56.440839 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:27:56 crc kubenswrapper[4818]: E1203 06:27:56.440857 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:56 crc kubenswrapper[4818]: E1203 06:27:56.440935 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:28:12.440906243 +0000 UTC m=+50.132515085 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.488420 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.488460 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.488479 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.488498 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.488512 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:56Z","lastTransitionTime":"2025-12-03T06:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.590895 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.590941 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.590950 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.590966 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.590975 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:56Z","lastTransitionTime":"2025-12-03T06:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.693237 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.693277 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.693288 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.693304 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.693316 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:56Z","lastTransitionTime":"2025-12-03T06:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.742341 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:27:56 crc kubenswrapper[4818]: E1203 06:27:56.742530 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.795130 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.795165 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.795174 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.795186 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.795194 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:56Z","lastTransitionTime":"2025-12-03T06:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.896773 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.896837 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.896849 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.896865 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.896876 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:56Z","lastTransitionTime":"2025-12-03T06:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.912317 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.923559 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.935445 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.942966 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovnkube-controller/0.log" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.945609 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerStarted","Data":"d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f"} Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.945736 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.947699 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.959138 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.971293 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.988695 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20eb93f400d5792232c343625af736d163a5f32f5a469e25feaf0af6de1446c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20eb93f400d5792232c343625af736d163a5f32f5a469e25feaf0af6de1446c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:27:55Z\\\",\\\"message\\\":\\\"or *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 06:27:55.468625 6107 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 06:27:55.468728 6107 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 06:27:55.469892 6107 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:27:55.469915 6107 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 06:27:55.469919 6107 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 06:27:55.469931 6107 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:27:55.469935 6107 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:27:55.469948 6107 factory.go:656] Stopping watch factory\\\\nI1203 06:27:55.469959 6107 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:27:55.469983 6107 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:27:55.469992 6107 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:27:55.469998 6107 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 06:27:55.470004 6107 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:56Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.999569 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.999598 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.999605 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.999618 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:56 crc kubenswrapper[4818]: I1203 06:27:56.999629 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:56Z","lastTransitionTime":"2025-12-03T06:27:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.002199 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.013177 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.028107 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.040060 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.053402 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.065341 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.081049 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.091387 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.102561 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.102603 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.102624 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.102640 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.102652 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:57Z","lastTransitionTime":"2025-12-03T06:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.103626 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.122202 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.134843 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.145370 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.155160 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.166106 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.184779 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.196275 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.204850 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.204893 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.204904 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.204921 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.204931 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:57Z","lastTransitionTime":"2025-12-03T06:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.207072 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.217531 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.227616 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.238580 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.248359 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.257268 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.275046 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20eb93f400d5792232c343625af736d163a5f32f5a469e25feaf0af6de1446c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:27:55Z\\\",\\\"message\\\":\\\"or *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 06:27:55.468625 6107 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 06:27:55.468728 6107 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 06:27:55.469892 6107 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:27:55.469915 6107 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 06:27:55.469919 6107 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 06:27:55.469931 6107 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:27:55.469935 6107 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:27:55.469948 6107 factory.go:656] Stopping watch factory\\\\nI1203 06:27:55.469959 6107 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:27:55.469983 6107 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:27:55.469992 6107 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:27:55.469998 6107 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 06:27:55.470004 6107 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.290619 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.307108 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.307136 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.307145 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.307158 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.307166 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:57Z","lastTransitionTime":"2025-12-03T06:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.409262 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.409310 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.409328 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.409350 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.409363 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:57Z","lastTransitionTime":"2025-12-03T06:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.511659 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.511693 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.511701 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.511716 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.511726 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:57Z","lastTransitionTime":"2025-12-03T06:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.614682 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.614717 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.614727 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.614743 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.614753 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:57Z","lastTransitionTime":"2025-12-03T06:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.718508 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.718577 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.718656 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.718683 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.718702 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:57Z","lastTransitionTime":"2025-12-03T06:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.736860 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.736871 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:57 crc kubenswrapper[4818]: E1203 06:27:57.737058 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:27:57 crc kubenswrapper[4818]: E1203 06:27:57.737276 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.822199 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.822266 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.822279 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.822300 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.822313 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:57Z","lastTransitionTime":"2025-12-03T06:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.925376 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.925420 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.925437 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.925455 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.925466 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:57Z","lastTransitionTime":"2025-12-03T06:27:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.951349 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovnkube-controller/1.log" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.952086 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovnkube-controller/0.log" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.955563 4818 generic.go:334] "Generic (PLEG): container finished" podID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerID="d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f" exitCode=1 Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.955607 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerDied","Data":"d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f"} Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.955642 4818 scope.go:117] "RemoveContainer" containerID="20eb93f400d5792232c343625af736d163a5f32f5a469e25feaf0af6de1446c9" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.957037 4818 scope.go:117] "RemoveContainer" containerID="d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f" Dec 03 06:27:57 crc kubenswrapper[4818]: E1203 06:27:57.957231 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" Dec 03 06:27:57 crc kubenswrapper[4818]: I1203 06:27:57.989889 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:57Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.008855 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.022045 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.032533 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.032588 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.032606 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.032630 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.032648 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:58Z","lastTransitionTime":"2025-12-03T06:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.038749 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.052542 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.074132 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.088050 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.103722 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.116153 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.131609 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.138070 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.138161 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.138180 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.138253 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.138278 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:58Z","lastTransitionTime":"2025-12-03T06:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.146802 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.157316 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.184958 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20eb93f400d5792232c343625af736d163a5f32f5a469e25feaf0af6de1446c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:27:55Z\\\",\\\"message\\\":\\\"or *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 06:27:55.468625 6107 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 06:27:55.468728 6107 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 06:27:55.469892 6107 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:27:55.469915 6107 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 06:27:55.469919 6107 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 06:27:55.469931 6107 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:27:55.469935 6107 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:27:55.469948 6107 factory.go:656] Stopping watch factory\\\\nI1203 06:27:55.469959 6107 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:27:55.469983 6107 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:27:55.469992 6107 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:27:55.469998 6107 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 06:27:55.470004 6107 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:27:57Z\\\",\\\"message\\\":\\\"208] Removed *v1.Pod event handler 6\\\\nI1203 06:27:57.132900 6238 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:27:57.132928 6238 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:27:57.132937 6238 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:27:57.132944 6238 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:27:57.134621 6238 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:27:57.134685 6238 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:27:57.134705 6238 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 06:27:57.134711 6238 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 06:27:57.134746 6238 factory.go:656] Stopping watch factory\\\\nI1203 06:27:57.134768 6238 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:27:57.134784 6238 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:27:57.134795 6238 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:27:57.134843 6238 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:27:57.134890 6238 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 06:27:57.134852 6238 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:27:57.135047 6238 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.200167 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.215034 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.240726 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.241035 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.241166 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.241284 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.241411 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:58Z","lastTransitionTime":"2025-12-03T06:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.344363 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.344425 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.344442 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.344465 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.344483 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:58Z","lastTransitionTime":"2025-12-03T06:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.447117 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.447164 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.447181 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.447204 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.447221 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:58Z","lastTransitionTime":"2025-12-03T06:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.550084 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.550137 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.550150 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.550169 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.550182 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:58Z","lastTransitionTime":"2025-12-03T06:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.653360 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.653454 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.653480 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.653517 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.653546 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:58Z","lastTransitionTime":"2025-12-03T06:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.737692 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:27:58 crc kubenswrapper[4818]: E1203 06:27:58.737924 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.757107 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.757167 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.757185 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.757208 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.757228 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:58Z","lastTransitionTime":"2025-12-03T06:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.860234 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.860325 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.860348 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.860379 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.860401 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:58Z","lastTransitionTime":"2025-12-03T06:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.867660 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc"] Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.868240 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.871559 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.872189 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.885655 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.902728 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.922658 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.934568 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.947169 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.963151 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.963225 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.963240 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.963269 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.963282 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:58Z","lastTransitionTime":"2025-12-03T06:27:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.964953 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.967037 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovnkube-controller/1.log" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.967055 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/57192c6e-16e0-458f-93aa-9af8af4eea19-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vr6bc\" (UID: \"57192c6e-16e0-458f-93aa-9af8af4eea19\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.967098 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/57192c6e-16e0-458f-93aa-9af8af4eea19-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vr6bc\" (UID: \"57192c6e-16e0-458f-93aa-9af8af4eea19\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.967245 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/57192c6e-16e0-458f-93aa-9af8af4eea19-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vr6bc\" (UID: \"57192c6e-16e0-458f-93aa-9af8af4eea19\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.967507 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2blz4\" (UniqueName: \"kubernetes.io/projected/57192c6e-16e0-458f-93aa-9af8af4eea19-kube-api-access-2blz4\") pod \"ovnkube-control-plane-749d76644c-vr6bc\" (UID: \"57192c6e-16e0-458f-93aa-9af8af4eea19\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.973300 4818 scope.go:117] "RemoveContainer" containerID="d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f" Dec 03 06:27:58 crc kubenswrapper[4818]: E1203 06:27:58.973450 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.980032 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:58 crc kubenswrapper[4818]: I1203 06:27:58.993908 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:58Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.016024 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20eb93f400d5792232c343625af736d163a5f32f5a469e25feaf0af6de1446c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:27:55Z\\\",\\\"message\\\":\\\"or *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 06:27:55.468625 6107 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 06:27:55.468728 6107 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 06:27:55.469892 6107 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:27:55.469915 6107 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 06:27:55.469919 6107 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 06:27:55.469931 6107 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:27:55.469935 6107 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:27:55.469948 6107 factory.go:656] Stopping watch factory\\\\nI1203 06:27:55.469959 6107 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:27:55.469983 6107 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:27:55.469992 6107 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:27:55.469998 6107 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 06:27:55.470004 6107 handler.go:208] Removed *v1.Node event handler 2\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:27:57Z\\\",\\\"message\\\":\\\"208] Removed *v1.Pod event handler 6\\\\nI1203 06:27:57.132900 6238 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:27:57.132928 6238 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:27:57.132937 6238 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:27:57.132944 6238 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:27:57.134621 6238 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:27:57.134685 6238 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:27:57.134705 6238 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 06:27:57.134711 6238 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 06:27:57.134746 6238 factory.go:656] Stopping watch factory\\\\nI1203 06:27:57.134768 6238 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:27:57.134784 6238 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:27:57.134795 6238 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:27:57.134843 6238 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:27:57.134890 6238 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 06:27:57.134852 6238 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:27:57.135047 6238 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.031568 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.050458 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.066132 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.066748 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.066847 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.066869 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.066893 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.066907 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:59Z","lastTransitionTime":"2025-12-03T06:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.068349 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/57192c6e-16e0-458f-93aa-9af8af4eea19-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vr6bc\" (UID: \"57192c6e-16e0-458f-93aa-9af8af4eea19\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.068412 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2blz4\" (UniqueName: \"kubernetes.io/projected/57192c6e-16e0-458f-93aa-9af8af4eea19-kube-api-access-2blz4\") pod \"ovnkube-control-plane-749d76644c-vr6bc\" (UID: \"57192c6e-16e0-458f-93aa-9af8af4eea19\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.068502 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/57192c6e-16e0-458f-93aa-9af8af4eea19-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vr6bc\" (UID: \"57192c6e-16e0-458f-93aa-9af8af4eea19\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.068533 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/57192c6e-16e0-458f-93aa-9af8af4eea19-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vr6bc\" (UID: \"57192c6e-16e0-458f-93aa-9af8af4eea19\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.069177 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/57192c6e-16e0-458f-93aa-9af8af4eea19-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vr6bc\" (UID: \"57192c6e-16e0-458f-93aa-9af8af4eea19\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.069856 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/57192c6e-16e0-458f-93aa-9af8af4eea19-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vr6bc\" (UID: \"57192c6e-16e0-458f-93aa-9af8af4eea19\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.073645 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/57192c6e-16e0-458f-93aa-9af8af4eea19-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vr6bc\" (UID: \"57192c6e-16e0-458f-93aa-9af8af4eea19\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.081113 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.086728 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2blz4\" (UniqueName: \"kubernetes.io/projected/57192c6e-16e0-458f-93aa-9af8af4eea19-kube-api-access-2blz4\") pod \"ovnkube-control-plane-749d76644c-vr6bc\" (UID: \"57192c6e-16e0-458f-93aa-9af8af4eea19\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.095036 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.109419 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.128352 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.144684 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.159630 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.169915 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.169974 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.169991 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.170012 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.170028 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:59Z","lastTransitionTime":"2025-12-03T06:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.175367 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.179051 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.192634 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: W1203 06:27:59.197228 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57192c6e_16e0_458f_93aa_9af8af4eea19.slice/crio-682b44428e81cc7049a981e37fae3bcd89231dbf12d55d8125e0470ac87dd7d2 WatchSource:0}: Error finding container 682b44428e81cc7049a981e37fae3bcd89231dbf12d55d8125e0470ac87dd7d2: Status 404 returned error can't find the container with id 682b44428e81cc7049a981e37fae3bcd89231dbf12d55d8125e0470ac87dd7d2 Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.212856 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.235513 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.255746 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.273555 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.273662 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.273704 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.273749 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.273764 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:59Z","lastTransitionTime":"2025-12-03T06:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.275101 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.292607 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.304372 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.317438 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.334486 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.349126 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.362736 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.376733 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.376782 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.376798 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.376847 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.376868 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:59Z","lastTransitionTime":"2025-12-03T06:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.386180 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:27:57Z\\\",\\\"message\\\":\\\"208] Removed *v1.Pod event handler 6\\\\nI1203 06:27:57.132900 6238 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:27:57.132928 6238 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:27:57.132937 6238 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:27:57.132944 6238 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:27:57.134621 6238 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:27:57.134685 6238 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:27:57.134705 6238 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 06:27:57.134711 6238 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 06:27:57.134746 6238 factory.go:656] Stopping watch factory\\\\nI1203 06:27:57.134768 6238 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:27:57.134784 6238 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:27:57.134795 6238 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:27:57.134843 6238 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:27:57.134890 6238 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 06:27:57.134852 6238 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:27:57.135047 6238 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.401970 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:27:59Z is after 2025-08-24T17:21:41Z" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.479203 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.479244 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.479255 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.479272 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.479285 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:59Z","lastTransitionTime":"2025-12-03T06:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.582867 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.583274 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.583413 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.583601 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.583796 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:59Z","lastTransitionTime":"2025-12-03T06:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.687146 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.687223 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.687242 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.687267 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.687284 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:59Z","lastTransitionTime":"2025-12-03T06:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.737524 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.737619 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:27:59 crc kubenswrapper[4818]: E1203 06:27:59.738033 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:27:59 crc kubenswrapper[4818]: E1203 06:27:59.738205 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.791087 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.791426 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.791568 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.791697 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.791895 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:59Z","lastTransitionTime":"2025-12-03T06:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.894865 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.894946 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.894965 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.894989 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.895008 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:59Z","lastTransitionTime":"2025-12-03T06:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.978946 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" event={"ID":"57192c6e-16e0-458f-93aa-9af8af4eea19","Type":"ContainerStarted","Data":"750933f24a9ca216d9efc07f0840656da96538ddfddf03931295e18ffbe79778"} Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.978999 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" event={"ID":"57192c6e-16e0-458f-93aa-9af8af4eea19","Type":"ContainerStarted","Data":"682b44428e81cc7049a981e37fae3bcd89231dbf12d55d8125e0470ac87dd7d2"} Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.997807 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.997875 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.997888 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.997904 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:27:59 crc kubenswrapper[4818]: I1203 06:27:59.997915 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:27:59Z","lastTransitionTime":"2025-12-03T06:27:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.100218 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.100278 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.100294 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.100313 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.100328 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:00Z","lastTransitionTime":"2025-12-03T06:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.202367 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.202393 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.202401 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.202413 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.202422 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:00Z","lastTransitionTime":"2025-12-03T06:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.305879 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.305932 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.305948 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.305971 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.305989 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:00Z","lastTransitionTime":"2025-12-03T06:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.389569 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-gq8r2"] Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.390461 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:00 crc kubenswrapper[4818]: E1203 06:28:00.390558 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.409449 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.409505 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.409523 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.409548 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.409565 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:00Z","lastTransitionTime":"2025-12-03T06:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.415776 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.435352 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.466265 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.483968 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs\") pod \"network-metrics-daemon-gq8r2\" (UID: \"2d8af736-67f3-4db0-a7a5-a24af4222b86\") " pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.484028 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fgxg\" (UniqueName: \"kubernetes.io/projected/2d8af736-67f3-4db0-a7a5-a24af4222b86-kube-api-access-9fgxg\") pod \"network-metrics-daemon-gq8r2\" (UID: \"2d8af736-67f3-4db0-a7a5-a24af4222b86\") " pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.487603 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.503454 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.512952 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.513021 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.513040 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.513063 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.513080 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:00Z","lastTransitionTime":"2025-12-03T06:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.520584 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.540474 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.559240 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.576439 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.585153 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs\") pod \"network-metrics-daemon-gq8r2\" (UID: \"2d8af736-67f3-4db0-a7a5-a24af4222b86\") " pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:00 crc kubenswrapper[4818]: E1203 06:28:00.585429 4818 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:28:00 crc kubenswrapper[4818]: E1203 06:28:00.585564 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs podName:2d8af736-67f3-4db0-a7a5-a24af4222b86 nodeName:}" failed. No retries permitted until 2025-12-03 06:28:01.085529945 +0000 UTC m=+38.777138777 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs") pod "network-metrics-daemon-gq8r2" (UID: "2d8af736-67f3-4db0-a7a5-a24af4222b86") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.585431 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fgxg\" (UniqueName: \"kubernetes.io/projected/2d8af736-67f3-4db0-a7a5-a24af4222b86-kube-api-access-9fgxg\") pod \"network-metrics-daemon-gq8r2\" (UID: \"2d8af736-67f3-4db0-a7a5-a24af4222b86\") " pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.598619 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.615214 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.617419 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.617466 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.617484 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.617508 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.617525 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:00Z","lastTransitionTime":"2025-12-03T06:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.618938 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fgxg\" (UniqueName: \"kubernetes.io/projected/2d8af736-67f3-4db0-a7a5-a24af4222b86-kube-api-access-9fgxg\") pod \"network-metrics-daemon-gq8r2\" (UID: \"2d8af736-67f3-4db0-a7a5-a24af4222b86\") " pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.635166 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.688002 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.709161 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.719292 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.719332 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.719348 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.719363 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.719375 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:00Z","lastTransitionTime":"2025-12-03T06:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.723219 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.736911 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:00 crc kubenswrapper[4818]: E1203 06:28:00.737156 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.750063 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:27:57Z\\\",\\\"message\\\":\\\"208] Removed *v1.Pod event handler 6\\\\nI1203 06:27:57.132900 6238 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:27:57.132928 6238 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:27:57.132937 6238 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:27:57.132944 6238 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:27:57.134621 6238 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:27:57.134685 6238 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:27:57.134705 6238 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 06:27:57.134711 6238 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 06:27:57.134746 6238 factory.go:656] Stopping watch factory\\\\nI1203 06:27:57.134768 6238 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:27:57.134784 6238 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:27:57.134795 6238 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:27:57.134843 6238 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:27:57.134890 6238 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 06:27:57.134852 6238 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:27:57.135047 6238 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.762349 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d8af736-67f3-4db0-a7a5-a24af4222b86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:28:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gq8r2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:00Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.821975 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.822026 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.822042 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.822066 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.822083 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:00Z","lastTransitionTime":"2025-12-03T06:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.925176 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.925213 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.925224 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.925239 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.925249 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:00Z","lastTransitionTime":"2025-12-03T06:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:00 crc kubenswrapper[4818]: I1203 06:28:00.985368 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" event={"ID":"57192c6e-16e0-458f-93aa-9af8af4eea19","Type":"ContainerStarted","Data":"25d96ab68b06a691db5cdfc715ad128e22902ba1d747a1b468be6283ae3579cb"} Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.006351 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:01Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.028857 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.028898 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.028915 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.028936 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.028955 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:01Z","lastTransitionTime":"2025-12-03T06:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.029927 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:27:57Z\\\",\\\"message\\\":\\\"208] Removed *v1.Pod event handler 6\\\\nI1203 06:27:57.132900 6238 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:27:57.132928 6238 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:27:57.132937 6238 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:27:57.132944 6238 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:27:57.134621 6238 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:27:57.134685 6238 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:27:57.134705 6238 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 06:27:57.134711 6238 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 06:27:57.134746 6238 factory.go:656] Stopping watch factory\\\\nI1203 06:27:57.134768 6238 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:27:57.134784 6238 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:27:57.134795 6238 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:27:57.134843 6238 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:27:57.134890 6238 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 06:27:57.134852 6238 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:27:57.135047 6238 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:01Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.048601 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d8af736-67f3-4db0-a7a5-a24af4222b86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:28:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gq8r2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:01Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.070138 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:01Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.086541 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:01Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.093678 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs\") pod \"network-metrics-daemon-gq8r2\" (UID: \"2d8af736-67f3-4db0-a7a5-a24af4222b86\") " pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:01 crc kubenswrapper[4818]: E1203 06:28:01.093848 4818 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:28:01 crc kubenswrapper[4818]: E1203 06:28:01.093920 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs podName:2d8af736-67f3-4db0-a7a5-a24af4222b86 nodeName:}" failed. No retries permitted until 2025-12-03 06:28:02.093898528 +0000 UTC m=+39.785507280 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs") pod "network-metrics-daemon-gq8r2" (UID: "2d8af736-67f3-4db0-a7a5-a24af4222b86") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.109988 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:01Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.128386 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:01Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.131634 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.131707 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.131738 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.131768 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.131794 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:01Z","lastTransitionTime":"2025-12-03T06:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.144997 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:01Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.164308 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:01Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.213096 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:01Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.233939 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.233986 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.233997 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.234014 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.234025 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:01Z","lastTransitionTime":"2025-12-03T06:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.243014 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:01Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.259508 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:01Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.272890 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:01Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.283046 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:01Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.293094 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://750933f24a9ca216d9efc07f0840656da96538ddfddf03931295e18ffbe79778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25d96ab68b06a691db5cdfc715ad128e22902ba1d747a1b468be6283ae3579cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:01Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.305567 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:01Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.317750 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:01Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.336290 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.336323 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.336335 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.336351 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.336364 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:01Z","lastTransitionTime":"2025-12-03T06:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.407331 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.409011 4818 scope.go:117] "RemoveContainer" containerID="d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f" Dec 03 06:28:01 crc kubenswrapper[4818]: E1203 06:28:01.409300 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.439907 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.439969 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.439988 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.440014 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.440034 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:01Z","lastTransitionTime":"2025-12-03T06:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.543170 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.543227 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.543239 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.543259 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.543273 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:01Z","lastTransitionTime":"2025-12-03T06:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.646357 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.646416 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.646426 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.646448 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.646461 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:01Z","lastTransitionTime":"2025-12-03T06:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.737179 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.737271 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.737347 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:01 crc kubenswrapper[4818]: E1203 06:28:01.737387 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:01 crc kubenswrapper[4818]: E1203 06:28:01.737497 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:01 crc kubenswrapper[4818]: E1203 06:28:01.737602 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.749914 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.749974 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.749986 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.750009 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.750022 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:01Z","lastTransitionTime":"2025-12-03T06:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.852993 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.853058 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.853075 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.853100 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.853117 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:01Z","lastTransitionTime":"2025-12-03T06:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.955586 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.955655 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.955676 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.955699 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:01 crc kubenswrapper[4818]: I1203 06:28:01.955717 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:01Z","lastTransitionTime":"2025-12-03T06:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.058257 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.058341 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.058361 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.058386 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.058404 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:02Z","lastTransitionTime":"2025-12-03T06:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.108110 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs\") pod \"network-metrics-daemon-gq8r2\" (UID: \"2d8af736-67f3-4db0-a7a5-a24af4222b86\") " pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:02 crc kubenswrapper[4818]: E1203 06:28:02.108478 4818 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:28:02 crc kubenswrapper[4818]: E1203 06:28:02.108671 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs podName:2d8af736-67f3-4db0-a7a5-a24af4222b86 nodeName:}" failed. No retries permitted until 2025-12-03 06:28:04.108638543 +0000 UTC m=+41.800247295 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs") pod "network-metrics-daemon-gq8r2" (UID: "2d8af736-67f3-4db0-a7a5-a24af4222b86") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.161380 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.161445 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.161477 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.161507 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.161529 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:02Z","lastTransitionTime":"2025-12-03T06:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.264793 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.264839 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.264848 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.264862 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.264872 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:02Z","lastTransitionTime":"2025-12-03T06:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.367904 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.367977 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.367994 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.368016 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.368033 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:02Z","lastTransitionTime":"2025-12-03T06:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.470540 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.470577 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.470586 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.470601 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.470611 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:02Z","lastTransitionTime":"2025-12-03T06:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.573234 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.573279 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.573296 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.573319 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.573336 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:02Z","lastTransitionTime":"2025-12-03T06:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.675915 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.675943 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.675952 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.675967 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.675978 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:02Z","lastTransitionTime":"2025-12-03T06:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.737306 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:02 crc kubenswrapper[4818]: E1203 06:28:02.737423 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.753167 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.765065 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.778895 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.778955 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.778969 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.778986 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.778999 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:02Z","lastTransitionTime":"2025-12-03T06:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.792899 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.810652 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.828489 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.842364 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.860186 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.879509 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.881641 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.881744 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.881767 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.881791 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.881805 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:02Z","lastTransitionTime":"2025-12-03T06:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.893113 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.915764 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.931397 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.943932 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://750933f24a9ca216d9efc07f0840656da96538ddfddf03931295e18ffbe79778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25d96ab68b06a691db5cdfc715ad128e22902ba1d747a1b468be6283ae3579cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.962899 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.978976 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.984495 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.984566 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.984584 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.984608 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.984629 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:02Z","lastTransitionTime":"2025-12-03T06:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:02 crc kubenswrapper[4818]: I1203 06:28:02.995737 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:02Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.014625 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:27:57Z\\\",\\\"message\\\":\\\"208] Removed *v1.Pod event handler 6\\\\nI1203 06:27:57.132900 6238 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:27:57.132928 6238 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:27:57.132937 6238 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:27:57.132944 6238 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:27:57.134621 6238 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:27:57.134685 6238 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:27:57.134705 6238 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 06:27:57.134711 6238 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 06:27:57.134746 6238 factory.go:656] Stopping watch factory\\\\nI1203 06:27:57.134768 6238 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:27:57.134784 6238 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:27:57.134795 6238 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:27:57.134843 6238 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:27:57.134890 6238 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 06:27:57.134852 6238 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:27:57.135047 6238 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:03Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.024431 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d8af736-67f3-4db0-a7a5-a24af4222b86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:28:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gq8r2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:03Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.087627 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.087671 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.087680 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.087694 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.087706 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:03Z","lastTransitionTime":"2025-12-03T06:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.190733 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.190871 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.190890 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.190915 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.190933 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:03Z","lastTransitionTime":"2025-12-03T06:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.293316 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.293363 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.293372 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.293387 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.293398 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:03Z","lastTransitionTime":"2025-12-03T06:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.396859 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.396901 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.396910 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.396924 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.396934 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:03Z","lastTransitionTime":"2025-12-03T06:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.500094 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.500182 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.500203 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.500237 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.500255 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:03Z","lastTransitionTime":"2025-12-03T06:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.603359 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.603632 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.603756 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.603966 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.604040 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:03Z","lastTransitionTime":"2025-12-03T06:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.706616 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.706669 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.706681 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.706700 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.706720 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:03Z","lastTransitionTime":"2025-12-03T06:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.737603 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.737647 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:03 crc kubenswrapper[4818]: E1203 06:28:03.737717 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:03 crc kubenswrapper[4818]: E1203 06:28:03.737913 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.737986 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:03 crc kubenswrapper[4818]: E1203 06:28:03.738116 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.809854 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.809886 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.809895 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.809913 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.809923 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:03Z","lastTransitionTime":"2025-12-03T06:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.913309 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.913366 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.913383 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.913406 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:03 crc kubenswrapper[4818]: I1203 06:28:03.913422 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:03Z","lastTransitionTime":"2025-12-03T06:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.016092 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.016150 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.016168 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.016194 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.016213 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:04Z","lastTransitionTime":"2025-12-03T06:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.119725 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.119788 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.119801 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.119840 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.119857 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:04Z","lastTransitionTime":"2025-12-03T06:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.133637 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs\") pod \"network-metrics-daemon-gq8r2\" (UID: \"2d8af736-67f3-4db0-a7a5-a24af4222b86\") " pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:04 crc kubenswrapper[4818]: E1203 06:28:04.134004 4818 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:28:04 crc kubenswrapper[4818]: E1203 06:28:04.134145 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs podName:2d8af736-67f3-4db0-a7a5-a24af4222b86 nodeName:}" failed. No retries permitted until 2025-12-03 06:28:08.134111837 +0000 UTC m=+45.825720809 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs") pod "network-metrics-daemon-gq8r2" (UID: "2d8af736-67f3-4db0-a7a5-a24af4222b86") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.161632 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.161771 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.161796 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.161877 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.161913 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:04Z","lastTransitionTime":"2025-12-03T06:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:04 crc kubenswrapper[4818]: E1203 06:28:04.182521 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:04Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.188061 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.188102 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.188119 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.188143 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.188160 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:04Z","lastTransitionTime":"2025-12-03T06:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:04 crc kubenswrapper[4818]: E1203 06:28:04.209763 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:04Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.214323 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.214394 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.214415 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.214438 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.214454 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:04Z","lastTransitionTime":"2025-12-03T06:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:04 crc kubenswrapper[4818]: E1203 06:28:04.235149 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:04Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.243753 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.243841 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.243875 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.243905 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.243924 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:04Z","lastTransitionTime":"2025-12-03T06:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:04 crc kubenswrapper[4818]: E1203 06:28:04.266110 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:04Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.272049 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.272103 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.272119 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.272142 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.272158 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:04Z","lastTransitionTime":"2025-12-03T06:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:04 crc kubenswrapper[4818]: E1203 06:28:04.287996 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:04Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:04 crc kubenswrapper[4818]: E1203 06:28:04.288571 4818 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.290591 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.290721 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.290861 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.290963 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.291083 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:04Z","lastTransitionTime":"2025-12-03T06:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.393487 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.393589 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.393613 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.393647 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.393670 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:04Z","lastTransitionTime":"2025-12-03T06:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.496585 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.496640 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.496659 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.496682 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.496699 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:04Z","lastTransitionTime":"2025-12-03T06:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.598546 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.598892 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.598993 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.599107 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.599171 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:04Z","lastTransitionTime":"2025-12-03T06:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.701741 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.701788 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.701805 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.701862 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.701883 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:04Z","lastTransitionTime":"2025-12-03T06:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.737180 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:04 crc kubenswrapper[4818]: E1203 06:28:04.737346 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.804533 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.804650 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.804662 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.804696 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.804711 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:04Z","lastTransitionTime":"2025-12-03T06:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.907336 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.907404 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.907429 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.907457 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:04 crc kubenswrapper[4818]: I1203 06:28:04.907479 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:04Z","lastTransitionTime":"2025-12-03T06:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.010869 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.011289 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.011459 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.012112 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.012360 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:05Z","lastTransitionTime":"2025-12-03T06:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.115238 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.115286 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.115297 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.115313 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.115325 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:05Z","lastTransitionTime":"2025-12-03T06:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.218165 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.218214 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.218227 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.218240 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.218250 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:05Z","lastTransitionTime":"2025-12-03T06:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.321849 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.321923 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.321938 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.321958 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.321975 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:05Z","lastTransitionTime":"2025-12-03T06:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.424353 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.424395 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.424409 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.424425 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.424437 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:05Z","lastTransitionTime":"2025-12-03T06:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.527577 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.527614 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.527625 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.527640 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.527651 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:05Z","lastTransitionTime":"2025-12-03T06:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.631139 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.631190 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.631209 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.631234 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.631253 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:05Z","lastTransitionTime":"2025-12-03T06:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.733886 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.733933 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.733952 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.733981 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.733998 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:05Z","lastTransitionTime":"2025-12-03T06:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.737440 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.737515 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.737554 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:05 crc kubenswrapper[4818]: E1203 06:28:05.737642 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:05 crc kubenswrapper[4818]: E1203 06:28:05.737894 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:05 crc kubenswrapper[4818]: E1203 06:28:05.738013 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.837710 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.837776 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.837793 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.837833 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.837848 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:05Z","lastTransitionTime":"2025-12-03T06:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.941086 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.941310 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.941343 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.941356 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:05 crc kubenswrapper[4818]: I1203 06:28:05.941365 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:05Z","lastTransitionTime":"2025-12-03T06:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.044479 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.044529 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.044543 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.044562 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.044575 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:06Z","lastTransitionTime":"2025-12-03T06:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.147034 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.147130 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.147166 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.147195 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.147219 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:06Z","lastTransitionTime":"2025-12-03T06:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.251088 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.251178 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.251195 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.251214 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.251232 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:06Z","lastTransitionTime":"2025-12-03T06:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.355140 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.355202 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.355219 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.355241 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.355259 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:06Z","lastTransitionTime":"2025-12-03T06:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.458849 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.458921 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.458944 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.458972 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.458995 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:06Z","lastTransitionTime":"2025-12-03T06:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.561559 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.561881 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.561962 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.562029 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.562095 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:06Z","lastTransitionTime":"2025-12-03T06:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.663924 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.663970 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.663981 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.663999 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.664011 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:06Z","lastTransitionTime":"2025-12-03T06:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.736864 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:06 crc kubenswrapper[4818]: E1203 06:28:06.737041 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.766441 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.766696 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.766769 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.766900 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.766984 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:06Z","lastTransitionTime":"2025-12-03T06:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.869220 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.869269 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.869280 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.869298 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.869311 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:06Z","lastTransitionTime":"2025-12-03T06:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.972374 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.972683 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.972865 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.972954 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:06 crc kubenswrapper[4818]: I1203 06:28:06.973070 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:06Z","lastTransitionTime":"2025-12-03T06:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.077293 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.077582 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.077668 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.077752 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.077862 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:07Z","lastTransitionTime":"2025-12-03T06:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.181081 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.181157 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.181176 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.181201 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.181218 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:07Z","lastTransitionTime":"2025-12-03T06:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.284084 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.284483 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.284550 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.284639 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.284715 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:07Z","lastTransitionTime":"2025-12-03T06:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.387602 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.387646 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.387659 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.387675 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.387686 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:07Z","lastTransitionTime":"2025-12-03T06:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.490763 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.490871 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.490896 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.490924 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.490946 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:07Z","lastTransitionTime":"2025-12-03T06:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.593641 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.593689 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.593701 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.593720 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.593732 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:07Z","lastTransitionTime":"2025-12-03T06:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.696221 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.696281 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.696293 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.696313 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.696324 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:07Z","lastTransitionTime":"2025-12-03T06:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.737356 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.737395 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.737526 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:07 crc kubenswrapper[4818]: E1203 06:28:07.737616 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:07 crc kubenswrapper[4818]: E1203 06:28:07.737699 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:07 crc kubenswrapper[4818]: E1203 06:28:07.737892 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.800512 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.800614 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.800630 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.800654 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.800669 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:07Z","lastTransitionTime":"2025-12-03T06:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.903347 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.903403 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.903416 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.903438 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:07 crc kubenswrapper[4818]: I1203 06:28:07.903464 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:07Z","lastTransitionTime":"2025-12-03T06:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.007368 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.007471 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.007489 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.007515 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.007533 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:08Z","lastTransitionTime":"2025-12-03T06:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.110361 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.110456 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.110467 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.110487 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.110499 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:08Z","lastTransitionTime":"2025-12-03T06:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.186414 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs\") pod \"network-metrics-daemon-gq8r2\" (UID: \"2d8af736-67f3-4db0-a7a5-a24af4222b86\") " pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:08 crc kubenswrapper[4818]: E1203 06:28:08.186581 4818 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:28:08 crc kubenswrapper[4818]: E1203 06:28:08.186644 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs podName:2d8af736-67f3-4db0-a7a5-a24af4222b86 nodeName:}" failed. No retries permitted until 2025-12-03 06:28:16.186624571 +0000 UTC m=+53.878233323 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs") pod "network-metrics-daemon-gq8r2" (UID: "2d8af736-67f3-4db0-a7a5-a24af4222b86") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.213315 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.213356 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.213365 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.213379 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.213389 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:08Z","lastTransitionTime":"2025-12-03T06:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.316333 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.316411 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.316426 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.316451 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.316471 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:08Z","lastTransitionTime":"2025-12-03T06:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.421131 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.421182 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.421197 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.421220 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.421232 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:08Z","lastTransitionTime":"2025-12-03T06:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.524348 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.524401 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.524417 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.524438 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.524455 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:08Z","lastTransitionTime":"2025-12-03T06:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.628111 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.628161 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.628175 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.628197 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.628221 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:08Z","lastTransitionTime":"2025-12-03T06:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.732397 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.733055 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.733145 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.733231 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.733386 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:08Z","lastTransitionTime":"2025-12-03T06:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.736880 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:08 crc kubenswrapper[4818]: E1203 06:28:08.737021 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.836512 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.836561 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.836574 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.836599 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.836613 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:08Z","lastTransitionTime":"2025-12-03T06:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.940484 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.940541 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.940560 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.940584 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:08 crc kubenswrapper[4818]: I1203 06:28:08.940603 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:08Z","lastTransitionTime":"2025-12-03T06:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.044174 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.044238 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.044252 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.044272 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.044285 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:09Z","lastTransitionTime":"2025-12-03T06:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.148190 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.148249 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.148268 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.148292 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.148311 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:09Z","lastTransitionTime":"2025-12-03T06:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.252203 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.253024 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.253069 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.253106 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.253126 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:09Z","lastTransitionTime":"2025-12-03T06:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.356333 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.356396 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.356413 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.356438 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.356458 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:09Z","lastTransitionTime":"2025-12-03T06:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.459864 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.459955 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.459969 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.459985 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.459998 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:09Z","lastTransitionTime":"2025-12-03T06:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.563159 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.563212 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.563233 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.563260 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.563282 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:09Z","lastTransitionTime":"2025-12-03T06:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.666163 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.666237 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.666264 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.666293 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.666313 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:09Z","lastTransitionTime":"2025-12-03T06:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.736860 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.736892 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.736878 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:09 crc kubenswrapper[4818]: E1203 06:28:09.737083 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:09 crc kubenswrapper[4818]: E1203 06:28:09.737151 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:09 crc kubenswrapper[4818]: E1203 06:28:09.737385 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.769838 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.769907 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.769923 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.769947 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.769965 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:09Z","lastTransitionTime":"2025-12-03T06:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.872954 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.873000 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.873010 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.873032 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.873044 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:09Z","lastTransitionTime":"2025-12-03T06:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.976588 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.976644 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.976656 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.976681 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:09 crc kubenswrapper[4818]: I1203 06:28:09.976697 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:09Z","lastTransitionTime":"2025-12-03T06:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.079976 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.080034 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.080045 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.080065 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.080078 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:10Z","lastTransitionTime":"2025-12-03T06:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.182864 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.182937 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.182952 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.182978 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.182993 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:10Z","lastTransitionTime":"2025-12-03T06:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.285854 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.286636 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.286723 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.286807 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.286907 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:10Z","lastTransitionTime":"2025-12-03T06:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.389418 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.389530 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.389616 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.389658 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.389685 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:10Z","lastTransitionTime":"2025-12-03T06:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.492237 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.492280 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.492292 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.492307 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.492318 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:10Z","lastTransitionTime":"2025-12-03T06:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.595077 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.595142 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.595155 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.595179 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.595194 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:10Z","lastTransitionTime":"2025-12-03T06:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.697798 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.697861 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.697880 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.697906 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.697921 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:10Z","lastTransitionTime":"2025-12-03T06:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.736924 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:10 crc kubenswrapper[4818]: E1203 06:28:10.737142 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.801196 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.801267 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.801277 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.801321 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.801334 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:10Z","lastTransitionTime":"2025-12-03T06:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.904647 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.904730 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.904757 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.904788 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:10 crc kubenswrapper[4818]: I1203 06:28:10.904805 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:10Z","lastTransitionTime":"2025-12-03T06:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.008121 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.008206 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.008225 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.008255 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.008280 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:11Z","lastTransitionTime":"2025-12-03T06:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.111179 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.111261 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.111282 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.111317 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.111345 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:11Z","lastTransitionTime":"2025-12-03T06:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.214681 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.214729 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.214741 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.214760 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.214774 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:11Z","lastTransitionTime":"2025-12-03T06:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.318396 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.318447 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.318459 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.318483 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.318497 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:11Z","lastTransitionTime":"2025-12-03T06:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.421654 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.421720 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.421731 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.421752 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.421765 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:11Z","lastTransitionTime":"2025-12-03T06:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.524513 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.524564 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.524573 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.524591 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.524602 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:11Z","lastTransitionTime":"2025-12-03T06:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.627373 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.627444 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.627456 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.627476 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.627487 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:11Z","lastTransitionTime":"2025-12-03T06:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.730193 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.730254 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.730270 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.730289 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.730299 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:11Z","lastTransitionTime":"2025-12-03T06:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.736584 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.736584 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:11 crc kubenswrapper[4818]: E1203 06:28:11.736740 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:11 crc kubenswrapper[4818]: E1203 06:28:11.736790 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.736587 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:11 crc kubenswrapper[4818]: E1203 06:28:11.736898 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.832951 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.833012 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.833031 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.833051 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.833062 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:11Z","lastTransitionTime":"2025-12-03T06:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.936111 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.936158 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.936172 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.936192 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:11 crc kubenswrapper[4818]: I1203 06:28:11.936207 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:11Z","lastTransitionTime":"2025-12-03T06:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.039166 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.039223 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.039236 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.039257 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.039275 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:12Z","lastTransitionTime":"2025-12-03T06:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.108510 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.123291 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.131462 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.144133 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.145532 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.145584 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.145610 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.145642 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.145668 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:12Z","lastTransitionTime":"2025-12-03T06:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.160185 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://750933f24a9ca216d9efc07f0840656da96538ddfddf03931295e18ffbe79778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25d96ab68b06a691db5cdfc715ad128e22902ba1d747a1b468be6283ae3579cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.176073 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.194067 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.208098 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.235017 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:27:57Z\\\",\\\"message\\\":\\\"208] Removed *v1.Pod event handler 6\\\\nI1203 06:27:57.132900 6238 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:27:57.132928 6238 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:27:57.132937 6238 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:27:57.132944 6238 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:27:57.134621 6238 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:27:57.134685 6238 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:27:57.134705 6238 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 06:27:57.134711 6238 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 06:27:57.134746 6238 factory.go:656] Stopping watch factory\\\\nI1203 06:27:57.134768 6238 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:27:57.134784 6238 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:27:57.134795 6238 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:27:57.134843 6238 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:27:57.134890 6238 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 06:27:57.134852 6238 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:27:57.135047 6238 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.248458 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d8af736-67f3-4db0-a7a5-a24af4222b86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:28:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gq8r2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.248728 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.248785 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.248800 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.248848 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.248871 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:12Z","lastTransitionTime":"2025-12-03T06:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.262653 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.283292 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.299678 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.312756 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.323486 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.335727 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.352479 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.352530 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.352544 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.352567 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.352582 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:12Z","lastTransitionTime":"2025-12-03T06:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.359731 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.379198 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.393714 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.436569 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:28:12 crc kubenswrapper[4818]: E1203 06:28:12.436906 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:28:44.436849145 +0000 UTC m=+82.128457917 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.455507 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.455548 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.455559 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.455576 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.455588 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:12Z","lastTransitionTime":"2025-12-03T06:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.538558 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:12 crc kubenswrapper[4818]: E1203 06:28:12.538739 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.538792 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:12 crc kubenswrapper[4818]: E1203 06:28:12.538882 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:28:44.538852766 +0000 UTC m=+82.230461518 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.538923 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.538986 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:12 crc kubenswrapper[4818]: E1203 06:28:12.539055 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:28:12 crc kubenswrapper[4818]: E1203 06:28:12.539078 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:28:12 crc kubenswrapper[4818]: E1203 06:28:12.539104 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:28:12 crc kubenswrapper[4818]: E1203 06:28:12.539120 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:28:12 crc kubenswrapper[4818]: E1203 06:28:12.539085 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:28:44.539076922 +0000 UTC m=+82.230685674 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:28:12 crc kubenswrapper[4818]: E1203 06:28:12.539078 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:28:12 crc kubenswrapper[4818]: E1203 06:28:12.539198 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:28:44.539178634 +0000 UTC m=+82.230787386 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:28:12 crc kubenswrapper[4818]: E1203 06:28:12.539251 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:28:12 crc kubenswrapper[4818]: E1203 06:28:12.539281 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:28:12 crc kubenswrapper[4818]: E1203 06:28:12.539402 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:28:44.53938533 +0000 UTC m=+82.230994092 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.558540 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.558958 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.559081 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.559223 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.559340 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:12Z","lastTransitionTime":"2025-12-03T06:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.662852 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.662920 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.662938 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.662961 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.662977 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:12Z","lastTransitionTime":"2025-12-03T06:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.737060 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:12 crc kubenswrapper[4818]: E1203 06:28:12.737288 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.751304 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d407a21-4cfd-4505-865b-55e30c39b4e8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8047009041ecc11795cad4ced8b06bc2a4810c42023810db409854dfcaae2275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5faa1cad79099f32db2b2895d0fb547031ee6dfa6c4a9e306f355b93e4c7c959\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a192429ab94e49222181fd50bb4e9597966a1b6adb38c8ebaf0eb31fcf6b8bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.765189 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.765227 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.765237 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.765258 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.765276 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:12Z","lastTransitionTime":"2025-12-03T06:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.773528 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.793595 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.811173 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.823555 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.836199 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://750933f24a9ca216d9efc07f0840656da96538ddfddf03931295e18ffbe79778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25d96ab68b06a691db5cdfc715ad128e22902ba1d747a1b468be6283ae3579cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.853115 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.865560 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.867344 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.867375 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.867387 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.867406 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.867419 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:12Z","lastTransitionTime":"2025-12-03T06:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.877673 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.895671 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:27:57Z\\\",\\\"message\\\":\\\"208] Removed *v1.Pod event handler 6\\\\nI1203 06:27:57.132900 6238 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:27:57.132928 6238 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:27:57.132937 6238 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:27:57.132944 6238 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:27:57.134621 6238 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:27:57.134685 6238 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:27:57.134705 6238 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 06:27:57.134711 6238 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 06:27:57.134746 6238 factory.go:656] Stopping watch factory\\\\nI1203 06:27:57.134768 6238 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:27:57.134784 6238 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:27:57.134795 6238 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:27:57.134843 6238 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:27:57.134890 6238 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 06:27:57.134852 6238 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:27:57.135047 6238 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.908730 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d8af736-67f3-4db0-a7a5-a24af4222b86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:28:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gq8r2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.931364 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.943685 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.963246 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.970709 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.970960 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.971050 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.971135 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.971220 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:12Z","lastTransitionTime":"2025-12-03T06:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.979017 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.990539 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:12 crc kubenswrapper[4818]: I1203 06:28:12.999546 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:12Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.011120 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:13Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.074569 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.074637 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.074654 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.074677 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.074691 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:13Z","lastTransitionTime":"2025-12-03T06:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.177801 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.177918 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.177938 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.177968 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.177988 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:13Z","lastTransitionTime":"2025-12-03T06:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.281260 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.281314 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.281323 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.281342 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.281353 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:13Z","lastTransitionTime":"2025-12-03T06:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.384695 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.384749 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.384773 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.384797 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.384810 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:13Z","lastTransitionTime":"2025-12-03T06:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.488017 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.488095 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.488117 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.488144 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.488169 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:13Z","lastTransitionTime":"2025-12-03T06:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.591443 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.591500 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.591510 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.591532 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.591544 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:13Z","lastTransitionTime":"2025-12-03T06:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.694491 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.694543 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.694555 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.694577 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.694588 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:13Z","lastTransitionTime":"2025-12-03T06:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.736928 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.737028 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:13 crc kubenswrapper[4818]: E1203 06:28:13.737119 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.737243 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:13 crc kubenswrapper[4818]: E1203 06:28:13.737267 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:13 crc kubenswrapper[4818]: E1203 06:28:13.737443 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.797143 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.797206 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.797215 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.797237 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.797251 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:13Z","lastTransitionTime":"2025-12-03T06:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.900396 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.900462 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.900472 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.900494 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:13 crc kubenswrapper[4818]: I1203 06:28:13.900509 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:13Z","lastTransitionTime":"2025-12-03T06:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.002744 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.002788 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.002803 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.002848 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.002860 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:14Z","lastTransitionTime":"2025-12-03T06:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.105928 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.105975 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.105991 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.106011 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.106042 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:14Z","lastTransitionTime":"2025-12-03T06:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.208571 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.208614 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.208625 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.208643 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.208654 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:14Z","lastTransitionTime":"2025-12-03T06:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.311891 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.311953 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.311964 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.311986 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.311999 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:14Z","lastTransitionTime":"2025-12-03T06:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.317323 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.317386 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.317406 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.317431 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.317453 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:14Z","lastTransitionTime":"2025-12-03T06:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:14 crc kubenswrapper[4818]: E1203 06:28:14.337579 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:14Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.343568 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.343617 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.343628 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.343647 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.343659 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:14Z","lastTransitionTime":"2025-12-03T06:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:14 crc kubenswrapper[4818]: E1203 06:28:14.359654 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:14Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.364848 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.364926 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.364954 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.364984 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.365008 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:14Z","lastTransitionTime":"2025-12-03T06:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:14 crc kubenswrapper[4818]: E1203 06:28:14.380507 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:14Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.385164 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.385240 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.385267 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.385301 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.385323 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:14Z","lastTransitionTime":"2025-12-03T06:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:14 crc kubenswrapper[4818]: E1203 06:28:14.400655 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:14Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.406034 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.406223 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.406297 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.406373 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.406442 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:14Z","lastTransitionTime":"2025-12-03T06:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:14 crc kubenswrapper[4818]: E1203 06:28:14.422750 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:14Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:14 crc kubenswrapper[4818]: E1203 06:28:14.422938 4818 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.428595 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.428927 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.428972 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.429005 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.429034 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:14Z","lastTransitionTime":"2025-12-03T06:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.533346 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.533748 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.533845 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.533948 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.534041 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:14Z","lastTransitionTime":"2025-12-03T06:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.637233 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.637602 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.637715 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.637851 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.637940 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:14Z","lastTransitionTime":"2025-12-03T06:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.737371 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:14 crc kubenswrapper[4818]: E1203 06:28:14.737606 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.741544 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.741595 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.741613 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.741638 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.741654 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:14Z","lastTransitionTime":"2025-12-03T06:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.844569 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.844614 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.844626 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.844646 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.844658 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:14Z","lastTransitionTime":"2025-12-03T06:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.948313 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.948361 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.948379 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.948492 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:14 crc kubenswrapper[4818]: I1203 06:28:14.948554 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:14Z","lastTransitionTime":"2025-12-03T06:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.051100 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.051146 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.051166 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.051189 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.051203 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:15Z","lastTransitionTime":"2025-12-03T06:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.155306 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.155383 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.155396 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.155420 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.155431 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:15Z","lastTransitionTime":"2025-12-03T06:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.259671 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.259720 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.259735 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.259755 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.259767 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:15Z","lastTransitionTime":"2025-12-03T06:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.362920 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.362967 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.362977 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.362996 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.363008 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:15Z","lastTransitionTime":"2025-12-03T06:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.466753 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.466843 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.466859 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.466887 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.466902 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:15Z","lastTransitionTime":"2025-12-03T06:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.569671 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.569726 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.569738 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.569762 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.569786 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:15Z","lastTransitionTime":"2025-12-03T06:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.673700 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.673758 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.673769 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.673849 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.673865 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:15Z","lastTransitionTime":"2025-12-03T06:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.736862 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.736937 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.736986 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:15 crc kubenswrapper[4818]: E1203 06:28:15.737126 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:15 crc kubenswrapper[4818]: E1203 06:28:15.737579 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:15 crc kubenswrapper[4818]: E1203 06:28:15.737782 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.737964 4818 scope.go:117] "RemoveContainer" containerID="d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.776470 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.776970 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.776983 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.777001 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.777014 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:15Z","lastTransitionTime":"2025-12-03T06:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.879841 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.879918 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.879932 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.879955 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.879971 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:15Z","lastTransitionTime":"2025-12-03T06:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.982545 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.982590 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.982599 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.982637 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:15 crc kubenswrapper[4818]: I1203 06:28:15.982651 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:15Z","lastTransitionTime":"2025-12-03T06:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.049071 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovnkube-controller/1.log" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.051074 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerStarted","Data":"fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff"} Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.052487 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.086061 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.086116 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.086127 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.086146 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.086157 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:16Z","lastTransitionTime":"2025-12-03T06:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.091637 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.109881 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.123871 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.136644 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.151779 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.167307 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d407a21-4cfd-4505-865b-55e30c39b4e8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8047009041ecc11795cad4ced8b06bc2a4810c42023810db409854dfcaae2275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5faa1cad79099f32db2b2895d0fb547031ee6dfa6c4a9e306f355b93e4c7c959\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a192429ab94e49222181fd50bb4e9597966a1b6adb38c8ebaf0eb31fcf6b8bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.185510 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.190442 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs\") pod \"network-metrics-daemon-gq8r2\" (UID: \"2d8af736-67f3-4db0-a7a5-a24af4222b86\") " pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:16 crc kubenswrapper[4818]: E1203 06:28:16.191828 4818 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:28:16 crc kubenswrapper[4818]: E1203 06:28:16.191916 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs podName:2d8af736-67f3-4db0-a7a5-a24af4222b86 nodeName:}" failed. No retries permitted until 2025-12-03 06:28:32.191890041 +0000 UTC m=+69.883498803 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs") pod "network-metrics-daemon-gq8r2" (UID: "2d8af736-67f3-4db0-a7a5-a24af4222b86") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.194281 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.194406 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.194421 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.194445 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.194465 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:16Z","lastTransitionTime":"2025-12-03T06:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.204851 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.221745 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.232895 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.244033 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://750933f24a9ca216d9efc07f0840656da96538ddfddf03931295e18ffbe79778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25d96ab68b06a691db5cdfc715ad128e22902ba1d747a1b468be6283ae3579cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.269756 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.284381 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.297593 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.298730 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.298794 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.298808 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.298850 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.298863 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:16Z","lastTransitionTime":"2025-12-03T06:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.342905 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:27:57Z\\\",\\\"message\\\":\\\"208] Removed *v1.Pod event handler 6\\\\nI1203 06:27:57.132900 6238 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:27:57.132928 6238 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:27:57.132937 6238 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:27:57.132944 6238 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:27:57.134621 6238 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:27:57.134685 6238 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:27:57.134705 6238 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 06:27:57.134711 6238 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 06:27:57.134746 6238 factory.go:656] Stopping watch factory\\\\nI1203 06:27:57.134768 6238 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:27:57.134784 6238 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:27:57.134795 6238 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:27:57.134843 6238 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:27:57.134890 6238 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 06:27:57.134852 6238 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:27:57.135047 6238 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.364330 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d8af736-67f3-4db0-a7a5-a24af4222b86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:28:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gq8r2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.399062 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.406990 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.407030 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.407042 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.407057 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.407068 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:16Z","lastTransitionTime":"2025-12-03T06:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.420131 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:16Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.509950 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.510050 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.510071 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.510096 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.510113 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:16Z","lastTransitionTime":"2025-12-03T06:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.614489 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.614581 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.614609 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.614641 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.614665 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:16Z","lastTransitionTime":"2025-12-03T06:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.717437 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.717514 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.717538 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.717569 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.717593 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:16Z","lastTransitionTime":"2025-12-03T06:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.736915 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:16 crc kubenswrapper[4818]: E1203 06:28:16.737079 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.820487 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.820545 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.820562 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.820584 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.820598 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:16Z","lastTransitionTime":"2025-12-03T06:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.922801 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.922912 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.922938 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.922970 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:16 crc kubenswrapper[4818]: I1203 06:28:16.922994 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:16Z","lastTransitionTime":"2025-12-03T06:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.025977 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.026098 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.026122 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.026157 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.026177 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:17Z","lastTransitionTime":"2025-12-03T06:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.129120 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.129180 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.129195 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.129217 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.129234 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:17Z","lastTransitionTime":"2025-12-03T06:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.232010 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.232059 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.232071 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.232089 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.232120 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:17Z","lastTransitionTime":"2025-12-03T06:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.335045 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.335101 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.335117 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.335145 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.335161 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:17Z","lastTransitionTime":"2025-12-03T06:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.438139 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.438193 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.438211 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.438233 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.438250 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:17Z","lastTransitionTime":"2025-12-03T06:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.542133 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.542208 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.542232 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.542261 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.542283 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:17Z","lastTransitionTime":"2025-12-03T06:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.645548 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.645606 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.645628 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.645656 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.645679 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:17Z","lastTransitionTime":"2025-12-03T06:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.736951 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.737080 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.736951 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:17 crc kubenswrapper[4818]: E1203 06:28:17.737215 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:17 crc kubenswrapper[4818]: E1203 06:28:17.737327 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:17 crc kubenswrapper[4818]: E1203 06:28:17.737487 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.749156 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.749204 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.749221 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.749244 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.749263 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:17Z","lastTransitionTime":"2025-12-03T06:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.852317 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.852382 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.852396 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.852416 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.852428 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:17Z","lastTransitionTime":"2025-12-03T06:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.956144 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.956461 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.956590 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.956738 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:17 crc kubenswrapper[4818]: I1203 06:28:17.956944 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:17Z","lastTransitionTime":"2025-12-03T06:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.059670 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.060126 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.060264 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.060398 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.060573 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:18Z","lastTransitionTime":"2025-12-03T06:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.063523 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovnkube-controller/2.log" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.065272 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovnkube-controller/1.log" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.072218 4818 generic.go:334] "Generic (PLEG): container finished" podID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerID="fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff" exitCode=1 Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.072302 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerDied","Data":"fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff"} Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.072362 4818 scope.go:117] "RemoveContainer" containerID="d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.073754 4818 scope.go:117] "RemoveContainer" containerID="fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff" Dec 03 06:28:18 crc kubenswrapper[4818]: E1203 06:28:18.074100 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.090294 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.102794 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.115278 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.127537 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://750933f24a9ca216d9efc07f0840656da96538ddfddf03931295e18ffbe79778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25d96ab68b06a691db5cdfc715ad128e22902ba1d747a1b468be6283ae3579cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.139257 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d407a21-4cfd-4505-865b-55e30c39b4e8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8047009041ecc11795cad4ced8b06bc2a4810c42023810db409854dfcaae2275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5faa1cad79099f32db2b2895d0fb547031ee6dfa6c4a9e306f355b93e4c7c959\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a192429ab94e49222181fd50bb4e9597966a1b6adb38c8ebaf0eb31fcf6b8bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.153141 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.164691 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.164719 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.164727 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.164741 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.164750 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:18Z","lastTransitionTime":"2025-12-03T06:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.166748 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.182659 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.205030 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:27:57Z\\\",\\\"message\\\":\\\"208] Removed *v1.Pod event handler 6\\\\nI1203 06:27:57.132900 6238 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:27:57.132928 6238 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:27:57.132937 6238 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:27:57.132944 6238 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:27:57.134621 6238 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:27:57.134685 6238 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:27:57.134705 6238 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 06:27:57.134711 6238 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 06:27:57.134746 6238 factory.go:656] Stopping watch factory\\\\nI1203 06:27:57.134768 6238 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:27:57.134784 6238 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:27:57.134795 6238 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:27:57.134843 6238 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:27:57.134890 6238 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 06:27:57.134852 6238 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:27:57.135047 6238 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:28:17Z\\\",\\\"message\\\":\\\"37375 6453 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:28:17.137393 6453 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:28:17.138535 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:28:17.138567 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:28:17.138575 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:28:17.138607 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:28:17.138627 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:28:17.138664 6453 factory.go:656] Stopping watch factory\\\\nI1203 06:28:17.138688 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:28:17.138729 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:28:17.138743 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:28:17.138751 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:28:17.138758 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:28:17.138767 6453 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:28:17.138781 6453 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:28:17.138948 6453 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.219504 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d8af736-67f3-4db0-a7a5-a24af4222b86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:28:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gq8r2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.237652 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.252674 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.266797 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.267724 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.267787 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.267806 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.267856 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.267881 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:18Z","lastTransitionTime":"2025-12-03T06:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.281495 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.295412 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.312658 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.339188 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.355298 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:18Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.371551 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.371786 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.371964 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.372111 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.372267 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:18Z","lastTransitionTime":"2025-12-03T06:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.475575 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.475612 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.475624 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.475637 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.475647 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:18Z","lastTransitionTime":"2025-12-03T06:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.579812 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.579927 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.579951 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.579979 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.580001 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:18Z","lastTransitionTime":"2025-12-03T06:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.683233 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.683282 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.683307 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.683327 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.683342 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:18Z","lastTransitionTime":"2025-12-03T06:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.737384 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:18 crc kubenswrapper[4818]: E1203 06:28:18.737598 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.785776 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.785833 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.785846 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.785861 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.785873 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:18Z","lastTransitionTime":"2025-12-03T06:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.888619 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.888661 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.888673 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.888690 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.888702 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:18Z","lastTransitionTime":"2025-12-03T06:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.991013 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.991084 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.991107 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.991136 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:18 crc kubenswrapper[4818]: I1203 06:28:18.991158 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:18Z","lastTransitionTime":"2025-12-03T06:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.078238 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovnkube-controller/2.log" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.093649 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.093912 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.093988 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.094058 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.094147 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:19Z","lastTransitionTime":"2025-12-03T06:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.197889 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.197956 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.197976 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.198001 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.198018 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:19Z","lastTransitionTime":"2025-12-03T06:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.301057 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.301094 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.301104 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.301120 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.301132 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:19Z","lastTransitionTime":"2025-12-03T06:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.403111 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.403154 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.403164 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.403180 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.403190 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:19Z","lastTransitionTime":"2025-12-03T06:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.507042 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.507088 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.507100 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.507121 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.507133 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:19Z","lastTransitionTime":"2025-12-03T06:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.610205 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.610251 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.610262 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.610278 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.610291 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:19Z","lastTransitionTime":"2025-12-03T06:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.714215 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.714257 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.714271 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.714285 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.714296 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:19Z","lastTransitionTime":"2025-12-03T06:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.737260 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.737290 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:19 crc kubenswrapper[4818]: E1203 06:28:19.737665 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.737388 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:19 crc kubenswrapper[4818]: E1203 06:28:19.738035 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:19 crc kubenswrapper[4818]: E1203 06:28:19.737866 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.818014 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.818535 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.818665 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.818796 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.818941 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:19Z","lastTransitionTime":"2025-12-03T06:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.923044 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.923486 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.923667 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.923796 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:19 crc kubenswrapper[4818]: I1203 06:28:19.923923 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:19Z","lastTransitionTime":"2025-12-03T06:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.027555 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.027602 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.027615 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.027632 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.027644 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:20Z","lastTransitionTime":"2025-12-03T06:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.130906 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.131444 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.131564 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.131685 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.131834 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:20Z","lastTransitionTime":"2025-12-03T06:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.235534 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.235609 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.235631 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.235660 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.235683 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:20Z","lastTransitionTime":"2025-12-03T06:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.338386 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.338451 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.338477 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.338510 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.338531 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:20Z","lastTransitionTime":"2025-12-03T06:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.442291 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.442436 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.442461 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.442490 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.442511 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:20Z","lastTransitionTime":"2025-12-03T06:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.545364 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.545421 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.545438 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.545460 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.545478 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:20Z","lastTransitionTime":"2025-12-03T06:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.648489 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.649258 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.649373 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.649484 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.649556 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:20Z","lastTransitionTime":"2025-12-03T06:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.736923 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:20 crc kubenswrapper[4818]: E1203 06:28:20.737096 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.752638 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.752688 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.752699 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.752747 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.752763 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:20Z","lastTransitionTime":"2025-12-03T06:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.855874 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.855936 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.855961 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.855984 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.856000 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:20Z","lastTransitionTime":"2025-12-03T06:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.959624 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.959924 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.960030 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.960156 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:20 crc kubenswrapper[4818]: I1203 06:28:20.960263 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:20Z","lastTransitionTime":"2025-12-03T06:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.062928 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.063493 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.063757 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.064157 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.064477 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:21Z","lastTransitionTime":"2025-12-03T06:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.167431 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.167701 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.167893 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.168162 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.168417 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:21Z","lastTransitionTime":"2025-12-03T06:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.271714 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.271771 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.271780 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.271793 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.271803 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:21Z","lastTransitionTime":"2025-12-03T06:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.374504 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.374563 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.374576 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.374594 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.374606 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:21Z","lastTransitionTime":"2025-12-03T06:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.477301 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.477351 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.477365 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.477380 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.477392 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:21Z","lastTransitionTime":"2025-12-03T06:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.580282 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.580340 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.580356 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.580375 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.580388 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:21Z","lastTransitionTime":"2025-12-03T06:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.683673 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.683724 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.683735 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.683751 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.683764 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:21Z","lastTransitionTime":"2025-12-03T06:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.737565 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.737612 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.737687 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:21 crc kubenswrapper[4818]: E1203 06:28:21.737756 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:21 crc kubenswrapper[4818]: E1203 06:28:21.737934 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:21 crc kubenswrapper[4818]: E1203 06:28:21.738061 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.786871 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.786950 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.786967 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.786987 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.787000 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:21Z","lastTransitionTime":"2025-12-03T06:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.889673 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.889771 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.889794 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.889873 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.889902 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:21Z","lastTransitionTime":"2025-12-03T06:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.993233 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.993299 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.993310 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.993326 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:21 crc kubenswrapper[4818]: I1203 06:28:21.993337 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:21Z","lastTransitionTime":"2025-12-03T06:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.095879 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.095943 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.095957 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.095981 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.095999 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:22Z","lastTransitionTime":"2025-12-03T06:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.200408 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.200463 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.200480 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.200501 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.200515 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:22Z","lastTransitionTime":"2025-12-03T06:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.303333 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.303758 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.304024 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.304231 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.304443 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:22Z","lastTransitionTime":"2025-12-03T06:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.407915 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.408380 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.408588 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.408855 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.409094 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:22Z","lastTransitionTime":"2025-12-03T06:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.511933 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.512234 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.512395 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.512475 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.512550 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:22Z","lastTransitionTime":"2025-12-03T06:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.614893 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.615241 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.615393 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.615483 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.615589 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:22Z","lastTransitionTime":"2025-12-03T06:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.718232 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.718274 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.718284 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.718306 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.718321 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:22Z","lastTransitionTime":"2025-12-03T06:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.736662 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:22 crc kubenswrapper[4818]: E1203 06:28:22.736940 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.769928 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.787008 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.803354 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.819474 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.825532 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.825575 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.825587 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.825607 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.825622 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:22Z","lastTransitionTime":"2025-12-03T06:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.834621 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.850455 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d407a21-4cfd-4505-865b-55e30c39b4e8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8047009041ecc11795cad4ced8b06bc2a4810c42023810db409854dfcaae2275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5faa1cad79099f32db2b2895d0fb547031ee6dfa6c4a9e306f355b93e4c7c959\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a192429ab94e49222181fd50bb4e9597966a1b6adb38c8ebaf0eb31fcf6b8bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.863687 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.879075 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.892432 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.905303 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.920873 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://750933f24a9ca216d9efc07f0840656da96538ddfddf03931295e18ffbe79778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25d96ab68b06a691db5cdfc715ad128e22902ba1d747a1b468be6283ae3579cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.928215 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.928280 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.928296 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.928318 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.928335 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:22Z","lastTransitionTime":"2025-12-03T06:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.937305 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.950591 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.963358 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.986995 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:27:57Z\\\",\\\"message\\\":\\\"208] Removed *v1.Pod event handler 6\\\\nI1203 06:27:57.132900 6238 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:27:57.132928 6238 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:27:57.132937 6238 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:27:57.132944 6238 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:27:57.134621 6238 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:27:57.134685 6238 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:27:57.134705 6238 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 06:27:57.134711 6238 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 06:27:57.134746 6238 factory.go:656] Stopping watch factory\\\\nI1203 06:27:57.134768 6238 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:27:57.134784 6238 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:27:57.134795 6238 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:27:57.134843 6238 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:27:57.134890 6238 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 06:27:57.134852 6238 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:27:57.135047 6238 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:28:17Z\\\",\\\"message\\\":\\\"37375 6453 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:28:17.137393 6453 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:28:17.138535 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:28:17.138567 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:28:17.138575 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:28:17.138607 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:28:17.138627 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:28:17.138664 6453 factory.go:656] Stopping watch factory\\\\nI1203 06:28:17.138688 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:28:17.138729 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:28:17.138743 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:28:17.138751 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:28:17.138758 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:28:17.138767 6453 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:28:17.138781 6453 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:28:17.138948 6453 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:22 crc kubenswrapper[4818]: I1203 06:28:22.999691 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d8af736-67f3-4db0-a7a5-a24af4222b86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:28:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gq8r2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:22Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.017417 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:23Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.031653 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.031719 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.031731 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.031755 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.031772 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:23Z","lastTransitionTime":"2025-12-03T06:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.034045 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:23Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.134112 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.134182 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.134201 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.134230 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.134288 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:23Z","lastTransitionTime":"2025-12-03T06:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.239039 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.239071 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.239082 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.239098 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.239108 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:23Z","lastTransitionTime":"2025-12-03T06:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.341566 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.341603 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.341613 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.341627 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.341658 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:23Z","lastTransitionTime":"2025-12-03T06:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.444217 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.444265 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.444276 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.444291 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.444302 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:23Z","lastTransitionTime":"2025-12-03T06:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.547660 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.547732 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.547753 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.547781 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.547802 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:23Z","lastTransitionTime":"2025-12-03T06:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.651539 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.651631 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.651653 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.651686 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.651710 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:23Z","lastTransitionTime":"2025-12-03T06:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.736633 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.736763 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.736891 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:23 crc kubenswrapper[4818]: E1203 06:28:23.736936 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:23 crc kubenswrapper[4818]: E1203 06:28:23.737189 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:23 crc kubenswrapper[4818]: E1203 06:28:23.737422 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.755448 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.755499 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.755510 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.755526 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.755537 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:23Z","lastTransitionTime":"2025-12-03T06:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.858291 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.858399 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.858420 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.858480 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.858499 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:23Z","lastTransitionTime":"2025-12-03T06:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.962546 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.962607 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.962626 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.962648 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:23 crc kubenswrapper[4818]: I1203 06:28:23.962667 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:23Z","lastTransitionTime":"2025-12-03T06:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.065933 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.065973 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.065983 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.066000 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.066012 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:24Z","lastTransitionTime":"2025-12-03T06:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.170265 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.170339 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.170361 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.170391 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.170414 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:24Z","lastTransitionTime":"2025-12-03T06:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.273861 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.273898 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.273914 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.273935 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.273953 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:24Z","lastTransitionTime":"2025-12-03T06:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.378004 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.378068 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.378090 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.378119 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.378142 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:24Z","lastTransitionTime":"2025-12-03T06:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.480601 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.480634 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.480643 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.480657 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.480667 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:24Z","lastTransitionTime":"2025-12-03T06:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.583514 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.583554 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.583564 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.583579 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.583591 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:24Z","lastTransitionTime":"2025-12-03T06:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.686081 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.686139 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.686157 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.686181 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.686199 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:24Z","lastTransitionTime":"2025-12-03T06:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.737341 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:24 crc kubenswrapper[4818]: E1203 06:28:24.737541 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.747939 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.747991 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.748015 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.748037 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.748056 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:24Z","lastTransitionTime":"2025-12-03T06:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:24 crc kubenswrapper[4818]: E1203 06:28:24.763173 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:24Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.767317 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.767379 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.767390 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.767411 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.767426 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:24Z","lastTransitionTime":"2025-12-03T06:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:24 crc kubenswrapper[4818]: E1203 06:28:24.781991 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:24Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.786447 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.786477 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.786486 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.786502 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.786512 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:24Z","lastTransitionTime":"2025-12-03T06:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:24 crc kubenswrapper[4818]: E1203 06:28:24.802496 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:24Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.807294 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.807341 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.807357 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.807384 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.807395 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:24Z","lastTransitionTime":"2025-12-03T06:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:24 crc kubenswrapper[4818]: E1203 06:28:24.828347 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:24Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.832653 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.832690 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.832704 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.832724 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.832738 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:24Z","lastTransitionTime":"2025-12-03T06:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:24 crc kubenswrapper[4818]: E1203 06:28:24.848554 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:24Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:24 crc kubenswrapper[4818]: E1203 06:28:24.848710 4818 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.850638 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.850696 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.850720 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.850755 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.850781 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:24Z","lastTransitionTime":"2025-12-03T06:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.953907 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.953990 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.954015 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.954045 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:24 crc kubenswrapper[4818]: I1203 06:28:24.954065 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:24Z","lastTransitionTime":"2025-12-03T06:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.056608 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.056640 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.056649 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.056662 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.056672 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:25Z","lastTransitionTime":"2025-12-03T06:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.159763 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.159800 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.159829 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.159847 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.159859 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:25Z","lastTransitionTime":"2025-12-03T06:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.262054 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.262092 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.262106 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.262128 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.262144 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:25Z","lastTransitionTime":"2025-12-03T06:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.365500 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.365559 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.365570 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.365614 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.365628 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:25Z","lastTransitionTime":"2025-12-03T06:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.469199 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.469263 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.469275 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.469301 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.469316 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:25Z","lastTransitionTime":"2025-12-03T06:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.572704 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.572780 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.572801 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.572864 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.572888 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:25Z","lastTransitionTime":"2025-12-03T06:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.676603 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.676672 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.676697 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.676728 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.676752 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:25Z","lastTransitionTime":"2025-12-03T06:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.736955 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.737009 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:25 crc kubenswrapper[4818]: E1203 06:28:25.737097 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.736954 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:25 crc kubenswrapper[4818]: E1203 06:28:25.737225 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:25 crc kubenswrapper[4818]: E1203 06:28:25.737311 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.779128 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.779175 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.779189 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.779210 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.779227 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:25Z","lastTransitionTime":"2025-12-03T06:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.882277 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.882349 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.882403 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.882428 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.882442 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:25Z","lastTransitionTime":"2025-12-03T06:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.984453 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.984524 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.984539 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.984557 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:25 crc kubenswrapper[4818]: I1203 06:28:25.984568 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:25Z","lastTransitionTime":"2025-12-03T06:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.087775 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.087929 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.087963 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.087987 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.088011 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:26Z","lastTransitionTime":"2025-12-03T06:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.190592 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.190634 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.190643 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.190659 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.190671 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:26Z","lastTransitionTime":"2025-12-03T06:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.294019 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.294080 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.294091 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.294113 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.294125 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:26Z","lastTransitionTime":"2025-12-03T06:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.398128 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.398183 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.398200 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.398218 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.398230 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:26Z","lastTransitionTime":"2025-12-03T06:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.500877 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.500932 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.500942 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.500958 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.500969 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:26Z","lastTransitionTime":"2025-12-03T06:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.603836 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.603895 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.603905 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.603924 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.603939 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:26Z","lastTransitionTime":"2025-12-03T06:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.706569 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.706627 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.706642 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.706669 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.706684 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:26Z","lastTransitionTime":"2025-12-03T06:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.736916 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:26 crc kubenswrapper[4818]: E1203 06:28:26.737129 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.809434 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.809484 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.809497 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.809516 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.809526 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:26Z","lastTransitionTime":"2025-12-03T06:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.911717 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.911799 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.911854 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.911908 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:26 crc kubenswrapper[4818]: I1203 06:28:26.911930 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:26Z","lastTransitionTime":"2025-12-03T06:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.014327 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.014392 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.014407 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.014434 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.014451 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:27Z","lastTransitionTime":"2025-12-03T06:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.116586 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.116638 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.116652 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.116670 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.116682 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:27Z","lastTransitionTime":"2025-12-03T06:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.219210 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.219281 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.219300 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.219844 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.219899 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:27Z","lastTransitionTime":"2025-12-03T06:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.323558 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.323610 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.323625 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.323646 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.323660 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:27Z","lastTransitionTime":"2025-12-03T06:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.426626 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.426677 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.426686 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.426709 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.426720 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:27Z","lastTransitionTime":"2025-12-03T06:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.533465 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.533519 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.533534 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.533555 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.533567 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:27Z","lastTransitionTime":"2025-12-03T06:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.636846 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.636917 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.636929 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.636953 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.636966 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:27Z","lastTransitionTime":"2025-12-03T06:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.737451 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.737536 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.737634 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:27 crc kubenswrapper[4818]: E1203 06:28:27.737681 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:27 crc kubenswrapper[4818]: E1203 06:28:27.737928 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:27 crc kubenswrapper[4818]: E1203 06:28:27.738020 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.739055 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.739100 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.739114 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.739132 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.739144 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:27Z","lastTransitionTime":"2025-12-03T06:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.841930 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.841968 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.841978 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.841992 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.842001 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:27Z","lastTransitionTime":"2025-12-03T06:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.945332 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.945378 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.945391 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.945414 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:27 crc kubenswrapper[4818]: I1203 06:28:27.945426 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:27Z","lastTransitionTime":"2025-12-03T06:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.048437 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.048490 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.048507 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.048530 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.048547 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:28Z","lastTransitionTime":"2025-12-03T06:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.151498 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.151541 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.151554 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.151573 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.151591 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:28Z","lastTransitionTime":"2025-12-03T06:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.254876 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.254951 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.254970 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.254999 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.255018 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:28Z","lastTransitionTime":"2025-12-03T06:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.357944 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.357993 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.358006 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.358024 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.358039 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:28Z","lastTransitionTime":"2025-12-03T06:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.460236 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.460290 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.460305 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.460329 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.460343 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:28Z","lastTransitionTime":"2025-12-03T06:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.563054 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.563110 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.563125 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.563146 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.563160 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:28Z","lastTransitionTime":"2025-12-03T06:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.666614 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.666683 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.666697 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.666717 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.666731 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:28Z","lastTransitionTime":"2025-12-03T06:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.737193 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:28 crc kubenswrapper[4818]: E1203 06:28:28.737409 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.769114 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.769167 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.769177 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.769198 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.769212 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:28Z","lastTransitionTime":"2025-12-03T06:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.871958 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.872021 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.872035 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.872060 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.872076 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:28Z","lastTransitionTime":"2025-12-03T06:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.975003 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.975049 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.975063 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.975080 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:28 crc kubenswrapper[4818]: I1203 06:28:28.975092 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:28Z","lastTransitionTime":"2025-12-03T06:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.079753 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.079801 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.079828 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.079848 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.079862 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:29Z","lastTransitionTime":"2025-12-03T06:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.182866 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.182929 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.182941 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.182963 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.182973 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:29Z","lastTransitionTime":"2025-12-03T06:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.289287 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.289352 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.289371 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.289401 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.289417 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:29Z","lastTransitionTime":"2025-12-03T06:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.392405 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.392453 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.392467 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.392486 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.392499 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:29Z","lastTransitionTime":"2025-12-03T06:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.495213 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.495255 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.495265 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.495281 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.495292 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:29Z","lastTransitionTime":"2025-12-03T06:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.598873 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.598951 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.598966 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.598991 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.599012 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:29Z","lastTransitionTime":"2025-12-03T06:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.702908 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.702956 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.702965 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.702982 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.702994 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:29Z","lastTransitionTime":"2025-12-03T06:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.737501 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.737571 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:29 crc kubenswrapper[4818]: E1203 06:28:29.737670 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.737571 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:29 crc kubenswrapper[4818]: E1203 06:28:29.737860 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:29 crc kubenswrapper[4818]: E1203 06:28:29.737931 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.806456 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.806506 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.806518 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.806537 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.806551 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:29Z","lastTransitionTime":"2025-12-03T06:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.910197 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.910249 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.910263 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.910288 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:29 crc kubenswrapper[4818]: I1203 06:28:29.910303 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:29Z","lastTransitionTime":"2025-12-03T06:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.013161 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.013214 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.013232 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.013254 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.013271 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:30Z","lastTransitionTime":"2025-12-03T06:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.116679 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.116740 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.116764 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.116795 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.116855 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:30Z","lastTransitionTime":"2025-12-03T06:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.220292 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.220379 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.220393 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.220415 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.220431 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:30Z","lastTransitionTime":"2025-12-03T06:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.323037 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.323153 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.323176 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.323200 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.323218 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:30Z","lastTransitionTime":"2025-12-03T06:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.426000 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.426084 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.426107 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.426177 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.426197 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:30Z","lastTransitionTime":"2025-12-03T06:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.529175 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.529218 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.529237 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.529262 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.529282 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:30Z","lastTransitionTime":"2025-12-03T06:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.631131 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.631171 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.631182 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.631195 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.631206 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:30Z","lastTransitionTime":"2025-12-03T06:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.734113 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.734143 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.734153 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.734186 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.734196 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:30Z","lastTransitionTime":"2025-12-03T06:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.737512 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:30 crc kubenswrapper[4818]: E1203 06:28:30.737600 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.837207 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.837256 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.837269 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.837290 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.837307 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:30Z","lastTransitionTime":"2025-12-03T06:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.939624 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.939664 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.939674 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.939689 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:30 crc kubenswrapper[4818]: I1203 06:28:30.939699 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:30Z","lastTransitionTime":"2025-12-03T06:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.042382 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.042464 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.042483 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.042506 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.042523 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:31Z","lastTransitionTime":"2025-12-03T06:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.146471 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.146520 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.146536 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.146560 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.146577 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:31Z","lastTransitionTime":"2025-12-03T06:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.249262 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.249297 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.249308 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.249324 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.249337 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:31Z","lastTransitionTime":"2025-12-03T06:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.351954 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.352012 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.352024 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.352039 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.352079 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:31Z","lastTransitionTime":"2025-12-03T06:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.454201 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.454233 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.454241 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.454254 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.454263 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:31Z","lastTransitionTime":"2025-12-03T06:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.556494 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.556578 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.556596 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.556617 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.556635 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:31Z","lastTransitionTime":"2025-12-03T06:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.660272 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.660309 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.660318 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.660334 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.660343 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:31Z","lastTransitionTime":"2025-12-03T06:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.737114 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.737178 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:31 crc kubenswrapper[4818]: E1203 06:28:31.737241 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.737185 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:31 crc kubenswrapper[4818]: E1203 06:28:31.737318 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:31 crc kubenswrapper[4818]: E1203 06:28:31.737365 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.763107 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.763139 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.763155 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.763174 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.763185 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:31Z","lastTransitionTime":"2025-12-03T06:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.865915 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.865965 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.865976 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.866007 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.866024 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:31Z","lastTransitionTime":"2025-12-03T06:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.967843 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.967883 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.967894 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.967910 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:31 crc kubenswrapper[4818]: I1203 06:28:31.967950 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:31Z","lastTransitionTime":"2025-12-03T06:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.070945 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.071003 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.071015 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.071036 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.071050 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:32Z","lastTransitionTime":"2025-12-03T06:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.173652 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.173710 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.173727 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.173755 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.173770 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:32Z","lastTransitionTime":"2025-12-03T06:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.274195 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs\") pod \"network-metrics-daemon-gq8r2\" (UID: \"2d8af736-67f3-4db0-a7a5-a24af4222b86\") " pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:32 crc kubenswrapper[4818]: E1203 06:28:32.274413 4818 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:28:32 crc kubenswrapper[4818]: E1203 06:28:32.274507 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs podName:2d8af736-67f3-4db0-a7a5-a24af4222b86 nodeName:}" failed. No retries permitted until 2025-12-03 06:29:04.274486066 +0000 UTC m=+101.966094818 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs") pod "network-metrics-daemon-gq8r2" (UID: "2d8af736-67f3-4db0-a7a5-a24af4222b86") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.276375 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.276428 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.276444 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.276465 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.276484 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:32Z","lastTransitionTime":"2025-12-03T06:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.379827 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.379868 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.379882 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.379898 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.379910 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:32Z","lastTransitionTime":"2025-12-03T06:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.482146 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.482177 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.482187 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.482202 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.482211 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:32Z","lastTransitionTime":"2025-12-03T06:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.585112 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.585152 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.585161 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.585176 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.585186 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:32Z","lastTransitionTime":"2025-12-03T06:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.687661 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.687719 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.687734 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.687759 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.687774 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:32Z","lastTransitionTime":"2025-12-03T06:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.740101 4818 scope.go:117] "RemoveContainer" containerID="fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff" Dec 03 06:28:32 crc kubenswrapper[4818]: E1203 06:28:32.740536 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.737224 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:32 crc kubenswrapper[4818]: E1203 06:28:32.741107 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.749075 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.761209 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://750933f24a9ca216d9efc07f0840656da96538ddfddf03931295e18ffbe79778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25d96ab68b06a691db5cdfc715ad128e22902ba1d747a1b468be6283ae3579cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.775103 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d407a21-4cfd-4505-865b-55e30c39b4e8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8047009041ecc11795cad4ced8b06bc2a4810c42023810db409854dfcaae2275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5faa1cad79099f32db2b2895d0fb547031ee6dfa6c4a9e306f355b93e4c7c959\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a192429ab94e49222181fd50bb4e9597966a1b6adb38c8ebaf0eb31fcf6b8bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.785867 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.789754 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.789855 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.789867 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.789889 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.789905 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:32Z","lastTransitionTime":"2025-12-03T06:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.797991 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.813051 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.836470 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d98eb1a0f9ac1b6d1d325ee7db2a815ceb18f53b5606e93b32cef47ffb14c03f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:27:57Z\\\",\\\"message\\\":\\\"208] Removed *v1.Pod event handler 6\\\\nI1203 06:27:57.132900 6238 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:27:57.132928 6238 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:27:57.132937 6238 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:27:57.132944 6238 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:27:57.134621 6238 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:27:57.134685 6238 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:27:57.134705 6238 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 06:27:57.134711 6238 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 06:27:57.134746 6238 factory.go:656] Stopping watch factory\\\\nI1203 06:27:57.134768 6238 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:27:57.134784 6238 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:27:57.134795 6238 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:27:57.134843 6238 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 06:27:57.134890 6238 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 06:27:57.134852 6238 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:27:57.135047 6238 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:28:17Z\\\",\\\"message\\\":\\\"37375 6453 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:28:17.137393 6453 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:28:17.138535 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:28:17.138567 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:28:17.138575 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:28:17.138607 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:28:17.138627 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:28:17.138664 6453 factory.go:656] Stopping watch factory\\\\nI1203 06:28:17.138688 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:28:17.138729 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:28:17.138743 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:28:17.138751 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:28:17.138758 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:28:17.138767 6453 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:28:17.138781 6453 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:28:17.138948 6453 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:28:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.848935 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d8af736-67f3-4db0-a7a5-a24af4222b86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:28:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gq8r2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.862409 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.877795 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.891278 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.892123 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.892175 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.892191 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.892210 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.892222 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:32Z","lastTransitionTime":"2025-12-03T06:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.905361 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.917469 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.929014 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.948093 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.960607 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.971599 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.980310 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.990258 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d407a21-4cfd-4505-865b-55e30c39b4e8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8047009041ecc11795cad4ced8b06bc2a4810c42023810db409854dfcaae2275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5faa1cad79099f32db2b2895d0fb547031ee6dfa6c4a9e306f355b93e4c7c959\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a192429ab94e49222181fd50bb4e9597966a1b6adb38c8ebaf0eb31fcf6b8bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.994979 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.995062 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.995104 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.995119 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:32 crc kubenswrapper[4818]: I1203 06:28:32.995128 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:32Z","lastTransitionTime":"2025-12-03T06:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.001892 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:32Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.013382 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.024764 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.040323 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.052314 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://750933f24a9ca216d9efc07f0840656da96538ddfddf03931295e18ffbe79778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25d96ab68b06a691db5cdfc715ad128e22902ba1d747a1b468be6283ae3579cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.066238 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.077868 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.091718 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.097894 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.097962 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.097974 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.098022 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.098040 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:33Z","lastTransitionTime":"2025-12-03T06:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.109751 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:28:17Z\\\",\\\"message\\\":\\\"37375 6453 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:28:17.137393 6453 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:28:17.138535 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:28:17.138567 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:28:17.138575 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:28:17.138607 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:28:17.138627 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:28:17.138664 6453 factory.go:656] Stopping watch factory\\\\nI1203 06:28:17.138688 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:28:17.138729 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:28:17.138743 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:28:17.138751 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:28:17.138758 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:28:17.138767 6453 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:28:17.138781 6453 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:28:17.138948 6453 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:28:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.120441 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d8af736-67f3-4db0-a7a5-a24af4222b86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:28:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gq8r2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.131964 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ctpzs_7b25b836-b639-4111-bcea-af7cc791ea32/kube-multus/0.log" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.132014 4818 generic.go:334] "Generic (PLEG): container finished" podID="7b25b836-b639-4111-bcea-af7cc791ea32" containerID="3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56" exitCode=1 Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.132054 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ctpzs" event={"ID":"7b25b836-b639-4111-bcea-af7cc791ea32","Type":"ContainerDied","Data":"3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56"} Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.132514 4818 scope.go:117] "RemoveContainer" containerID="3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.135977 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.147971 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.173189 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.190153 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.200929 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.200978 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.200988 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.201007 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.201016 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:33Z","lastTransitionTime":"2025-12-03T06:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.204849 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.217247 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.231377 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.252040 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.264514 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.275357 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.288785 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:28:32Z\\\",\\\"message\\\":\\\"2025-12-03T06:27:47+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a1c583e8-fae6-4b94-af4c-e7b26ade3559\\\\n2025-12-03T06:27:47+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a1c583e8-fae6-4b94-af4c-e7b26ade3559 to /host/opt/cni/bin/\\\\n2025-12-03T06:27:47Z [verbose] multus-daemon started\\\\n2025-12-03T06:27:47Z [verbose] Readiness Indicator file check\\\\n2025-12-03T06:28:32Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.303831 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.303932 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.303949 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.303968 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.303979 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:33Z","lastTransitionTime":"2025-12-03T06:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.308917 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.322989 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.335699 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.350334 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.362082 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.377739 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://750933f24a9ca216d9efc07f0840656da96538ddfddf03931295e18ffbe79778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25d96ab68b06a691db5cdfc715ad128e22902ba1d747a1b468be6283ae3579cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.392619 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d407a21-4cfd-4505-865b-55e30c39b4e8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8047009041ecc11795cad4ced8b06bc2a4810c42023810db409854dfcaae2275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5faa1cad79099f32db2b2895d0fb547031ee6dfa6c4a9e306f355b93e4c7c959\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a192429ab94e49222181fd50bb4e9597966a1b6adb38c8ebaf0eb31fcf6b8bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.406631 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.406672 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.406681 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.406699 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.406711 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:33Z","lastTransitionTime":"2025-12-03T06:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.411424 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.427027 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.439571 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.458455 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:28:17Z\\\",\\\"message\\\":\\\"37375 6453 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:28:17.137393 6453 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:28:17.138535 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:28:17.138567 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:28:17.138575 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:28:17.138607 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:28:17.138627 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:28:17.138664 6453 factory.go:656] Stopping watch factory\\\\nI1203 06:28:17.138688 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:28:17.138729 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:28:17.138743 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:28:17.138751 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:28:17.138758 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:28:17.138767 6453 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:28:17.138781 6453 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:28:17.138948 6453 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:28:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.471310 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d8af736-67f3-4db0-a7a5-a24af4222b86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:28:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gq8r2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.486351 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.498700 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:33Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.509306 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.509338 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.509348 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.509362 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.509371 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:33Z","lastTransitionTime":"2025-12-03T06:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.612490 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.612542 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.612554 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.612569 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.612917 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:33Z","lastTransitionTime":"2025-12-03T06:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.715166 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.715223 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.715238 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.715260 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.715496 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:33Z","lastTransitionTime":"2025-12-03T06:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.736948 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.736970 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.737014 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:33 crc kubenswrapper[4818]: E1203 06:28:33.737116 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:33 crc kubenswrapper[4818]: E1203 06:28:33.737233 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:33 crc kubenswrapper[4818]: E1203 06:28:33.737333 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.818611 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.818695 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.818712 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.818739 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.818755 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:33Z","lastTransitionTime":"2025-12-03T06:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.921193 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.921458 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.921607 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.921785 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:33 crc kubenswrapper[4818]: I1203 06:28:33.921934 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:33Z","lastTransitionTime":"2025-12-03T06:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.024686 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.025602 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.025767 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.025962 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.026114 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:34Z","lastTransitionTime":"2025-12-03T06:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.128984 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.129339 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.129569 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.129870 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.130107 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:34Z","lastTransitionTime":"2025-12-03T06:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.138113 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ctpzs_7b25b836-b639-4111-bcea-af7cc791ea32/kube-multus/0.log" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.138185 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ctpzs" event={"ID":"7b25b836-b639-4111-bcea-af7cc791ea32","Type":"ContainerStarted","Data":"8b92f0662c82eef1fb0023107da93fead648febd6e3880b77ad264ced57de7be"} Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.152165 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.177533 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:28:17Z\\\",\\\"message\\\":\\\"37375 6453 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:28:17.137393 6453 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:28:17.138535 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:28:17.138567 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:28:17.138575 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:28:17.138607 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:28:17.138627 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:28:17.138664 6453 factory.go:656] Stopping watch factory\\\\nI1203 06:28:17.138688 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:28:17.138729 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:28:17.138743 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:28:17.138751 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:28:17.138758 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:28:17.138767 6453 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:28:17.138781 6453 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:28:17.138948 6453 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:28:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.187106 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d8af736-67f3-4db0-a7a5-a24af4222b86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:28:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gq8r2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.199372 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.211982 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.227367 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.232497 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.232521 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.232531 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.232545 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.232555 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:34Z","lastTransitionTime":"2025-12-03T06:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.239451 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.280369 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.297518 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b92f0662c82eef1fb0023107da93fead648febd6e3880b77ad264ced57de7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:28:32Z\\\",\\\"message\\\":\\\"2025-12-03T06:27:47+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a1c583e8-fae6-4b94-af4c-e7b26ade3559\\\\n2025-12-03T06:27:47+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a1c583e8-fae6-4b94-af4c-e7b26ade3559 to /host/opt/cni/bin/\\\\n2025-12-03T06:27:47Z [verbose] multus-daemon started\\\\n2025-12-03T06:27:47Z [verbose] Readiness Indicator file check\\\\n2025-12-03T06:28:32Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.317192 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.329239 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.334524 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.334564 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.334575 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.334592 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.334605 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:34Z","lastTransitionTime":"2025-12-03T06:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.342381 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.355373 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.364620 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.376588 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://750933f24a9ca216d9efc07f0840656da96538ddfddf03931295e18ffbe79778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25d96ab68b06a691db5cdfc715ad128e22902ba1d747a1b468be6283ae3579cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.389717 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d407a21-4cfd-4505-865b-55e30c39b4e8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8047009041ecc11795cad4ced8b06bc2a4810c42023810db409854dfcaae2275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5faa1cad79099f32db2b2895d0fb547031ee6dfa6c4a9e306f355b93e4c7c959\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a192429ab94e49222181fd50bb4e9597966a1b6adb38c8ebaf0eb31fcf6b8bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.403561 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.417666 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:34Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.437457 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.437495 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.437506 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.437520 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.437530 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:34Z","lastTransitionTime":"2025-12-03T06:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.540114 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.540156 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.540170 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.540188 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.540201 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:34Z","lastTransitionTime":"2025-12-03T06:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.643493 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.643551 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.643568 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.643584 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.643595 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:34Z","lastTransitionTime":"2025-12-03T06:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.737328 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:34 crc kubenswrapper[4818]: E1203 06:28:34.737465 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.745564 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.745596 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.745605 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.745619 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.745629 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:34Z","lastTransitionTime":"2025-12-03T06:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.848692 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.848774 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.848795 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.848872 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.848908 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:34Z","lastTransitionTime":"2025-12-03T06:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.952086 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.952136 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.952148 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.952163 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:34 crc kubenswrapper[4818]: I1203 06:28:34.952174 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:34Z","lastTransitionTime":"2025-12-03T06:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.011215 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.011273 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.011291 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.011316 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.011333 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:35Z","lastTransitionTime":"2025-12-03T06:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:35 crc kubenswrapper[4818]: E1203 06:28:35.028657 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.032656 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.032724 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.032746 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.032772 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.032793 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:35Z","lastTransitionTime":"2025-12-03T06:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:35 crc kubenswrapper[4818]: E1203 06:28:35.046352 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.049599 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.049685 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.049703 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.049724 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.049740 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:35Z","lastTransitionTime":"2025-12-03T06:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:35 crc kubenswrapper[4818]: E1203 06:28:35.063322 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.066929 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.067016 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.067031 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.067048 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.067060 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:35Z","lastTransitionTime":"2025-12-03T06:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:35 crc kubenswrapper[4818]: E1203 06:28:35.081909 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.085999 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.086044 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.086054 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.086071 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.086082 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:35Z","lastTransitionTime":"2025-12-03T06:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:35 crc kubenswrapper[4818]: E1203 06:28:35.102892 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:35Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:35 crc kubenswrapper[4818]: E1203 06:28:35.103049 4818 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.104720 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.104754 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.104765 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.104779 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.104788 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:35Z","lastTransitionTime":"2025-12-03T06:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.206790 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.206897 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.206907 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.206922 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.206933 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:35Z","lastTransitionTime":"2025-12-03T06:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.309839 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.309876 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.309885 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.309901 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.309910 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:35Z","lastTransitionTime":"2025-12-03T06:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.412808 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.412878 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.412888 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.412903 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.412914 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:35Z","lastTransitionTime":"2025-12-03T06:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.516251 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.516304 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.516316 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.516335 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.516350 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:35Z","lastTransitionTime":"2025-12-03T06:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.619109 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.619211 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.619244 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.619274 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.619294 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:35Z","lastTransitionTime":"2025-12-03T06:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.722190 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.722227 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.722236 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.722248 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.722260 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:35Z","lastTransitionTime":"2025-12-03T06:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.736514 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.736538 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.736602 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:35 crc kubenswrapper[4818]: E1203 06:28:35.736718 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:35 crc kubenswrapper[4818]: E1203 06:28:35.736831 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:35 crc kubenswrapper[4818]: E1203 06:28:35.736980 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.824987 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.825030 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.825040 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.825056 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.825068 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:35Z","lastTransitionTime":"2025-12-03T06:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.927459 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.927499 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.927508 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.927523 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:35 crc kubenswrapper[4818]: I1203 06:28:35.927532 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:35Z","lastTransitionTime":"2025-12-03T06:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.030642 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.030683 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.030695 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.030720 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.030730 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:36Z","lastTransitionTime":"2025-12-03T06:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.133351 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.133389 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.133400 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.133416 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.133430 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:36Z","lastTransitionTime":"2025-12-03T06:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.235509 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.235548 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.235558 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.235572 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.235583 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:36Z","lastTransitionTime":"2025-12-03T06:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.338133 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.338181 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.338194 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.338215 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.338229 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:36Z","lastTransitionTime":"2025-12-03T06:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.440889 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.440983 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.441002 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.441026 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.441046 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:36Z","lastTransitionTime":"2025-12-03T06:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.543222 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.543264 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.543272 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.543284 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.543294 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:36Z","lastTransitionTime":"2025-12-03T06:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.647177 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.647263 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.647282 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.647304 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.647324 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:36Z","lastTransitionTime":"2025-12-03T06:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.738217 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:36 crc kubenswrapper[4818]: E1203 06:28:36.738442 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.749440 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.749494 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.749510 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.749531 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.749549 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:36Z","lastTransitionTime":"2025-12-03T06:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.852228 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.852296 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.852318 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.852345 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.852366 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:36Z","lastTransitionTime":"2025-12-03T06:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.954729 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.955093 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.955223 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.955369 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:36 crc kubenswrapper[4818]: I1203 06:28:36.955501 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:36Z","lastTransitionTime":"2025-12-03T06:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.058612 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.058673 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.058697 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.058714 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.058725 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:37Z","lastTransitionTime":"2025-12-03T06:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.160476 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.160508 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.160516 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.160529 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.160540 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:37Z","lastTransitionTime":"2025-12-03T06:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.263983 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.264050 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.264064 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.264081 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.264095 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:37Z","lastTransitionTime":"2025-12-03T06:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.366543 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.366592 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.366603 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.366621 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.366634 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:37Z","lastTransitionTime":"2025-12-03T06:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.469273 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.469575 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.469659 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.469771 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.469881 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:37Z","lastTransitionTime":"2025-12-03T06:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.572643 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.573112 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.573334 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.573541 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.573744 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:37Z","lastTransitionTime":"2025-12-03T06:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.676884 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.676940 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.676957 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.676981 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.676998 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:37Z","lastTransitionTime":"2025-12-03T06:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.737364 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.737433 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:37 crc kubenswrapper[4818]: E1203 06:28:37.737546 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.737374 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:37 crc kubenswrapper[4818]: E1203 06:28:37.737733 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:37 crc kubenswrapper[4818]: E1203 06:28:37.737873 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.779478 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.779521 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.779533 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.779549 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.779561 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:37Z","lastTransitionTime":"2025-12-03T06:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.881973 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.882043 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.882061 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.882084 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.882104 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:37Z","lastTransitionTime":"2025-12-03T06:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.984714 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.984775 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.984798 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.984859 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:37 crc kubenswrapper[4818]: I1203 06:28:37.984882 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:37Z","lastTransitionTime":"2025-12-03T06:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.088530 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.089699 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.089907 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.090084 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.090215 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:38Z","lastTransitionTime":"2025-12-03T06:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.194013 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.194085 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.194108 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.194136 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.194157 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:38Z","lastTransitionTime":"2025-12-03T06:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.297583 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.297629 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.297645 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.297668 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.297684 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:38Z","lastTransitionTime":"2025-12-03T06:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.400893 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.401003 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.401017 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.401035 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.401048 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:38Z","lastTransitionTime":"2025-12-03T06:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.504223 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.504295 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.504312 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.504337 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.504354 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:38Z","lastTransitionTime":"2025-12-03T06:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.607338 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.607388 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.607399 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.607417 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.607430 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:38Z","lastTransitionTime":"2025-12-03T06:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.710485 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.710528 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.710539 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.710555 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.710566 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:38Z","lastTransitionTime":"2025-12-03T06:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.737073 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:38 crc kubenswrapper[4818]: E1203 06:28:38.737277 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.813354 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.813434 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.813452 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.813477 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.813494 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:38Z","lastTransitionTime":"2025-12-03T06:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.916730 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.917032 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.917131 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.917240 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:38 crc kubenswrapper[4818]: I1203 06:28:38.917330 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:38Z","lastTransitionTime":"2025-12-03T06:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.020292 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.020355 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.020373 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.020396 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.020416 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:39Z","lastTransitionTime":"2025-12-03T06:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.123039 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.123129 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.123153 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.123207 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.123232 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:39Z","lastTransitionTime":"2025-12-03T06:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.225420 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.225570 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.225607 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.225621 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.225631 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:39Z","lastTransitionTime":"2025-12-03T06:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.328549 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.328596 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.328607 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.328622 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.328632 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:39Z","lastTransitionTime":"2025-12-03T06:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.432145 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.432196 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.432211 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.432232 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.432247 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:39Z","lastTransitionTime":"2025-12-03T06:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.534350 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.534410 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.534422 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.534440 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.534452 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:39Z","lastTransitionTime":"2025-12-03T06:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.638145 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.638213 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.638230 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.638255 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.638274 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:39Z","lastTransitionTime":"2025-12-03T06:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.736750 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.736747 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.736751 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:39 crc kubenswrapper[4818]: E1203 06:28:39.736956 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:39 crc kubenswrapper[4818]: E1203 06:28:39.737110 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:39 crc kubenswrapper[4818]: E1203 06:28:39.737155 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.740998 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.741063 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.741085 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.741111 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.741135 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:39Z","lastTransitionTime":"2025-12-03T06:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.843594 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.843634 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.843643 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.843658 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.843668 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:39Z","lastTransitionTime":"2025-12-03T06:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.946634 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.946677 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.946689 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.946705 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:39 crc kubenswrapper[4818]: I1203 06:28:39.946716 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:39Z","lastTransitionTime":"2025-12-03T06:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.049157 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.049214 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.049232 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.049255 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.049272 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:40Z","lastTransitionTime":"2025-12-03T06:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.152036 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.152111 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.152134 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.152164 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.152186 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:40Z","lastTransitionTime":"2025-12-03T06:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.255372 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.255420 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.255430 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.255445 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.255456 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:40Z","lastTransitionTime":"2025-12-03T06:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.358955 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.358989 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.359000 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.359033 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.359045 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:40Z","lastTransitionTime":"2025-12-03T06:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.463402 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.463470 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.463494 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.463566 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.463591 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:40Z","lastTransitionTime":"2025-12-03T06:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.565796 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.566050 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.566159 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.566297 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.566377 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:40Z","lastTransitionTime":"2025-12-03T06:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.669299 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.669334 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.669344 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.669359 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.669370 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:40Z","lastTransitionTime":"2025-12-03T06:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.737725 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:40 crc kubenswrapper[4818]: E1203 06:28:40.737950 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.771811 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.771900 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.771938 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.771976 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.772000 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:40Z","lastTransitionTime":"2025-12-03T06:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.875333 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.875387 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.875403 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.875427 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.875446 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:40Z","lastTransitionTime":"2025-12-03T06:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.977608 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.977754 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.977778 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.977807 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:40 crc kubenswrapper[4818]: I1203 06:28:40.977857 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:40Z","lastTransitionTime":"2025-12-03T06:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.080109 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.080148 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.080157 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.080170 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.080178 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:41Z","lastTransitionTime":"2025-12-03T06:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.182728 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.183745 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.183897 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.184051 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.184179 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:41Z","lastTransitionTime":"2025-12-03T06:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.287171 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.287747 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.288109 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.288240 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.288363 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:41Z","lastTransitionTime":"2025-12-03T06:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.393030 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.393451 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.393618 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.393796 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.394014 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:41Z","lastTransitionTime":"2025-12-03T06:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.496407 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.496703 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.496932 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.497229 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.497621 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:41Z","lastTransitionTime":"2025-12-03T06:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.600786 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.601172 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.601310 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.601448 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.601583 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:41Z","lastTransitionTime":"2025-12-03T06:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.705902 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.705951 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.706150 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.706173 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.706190 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:41Z","lastTransitionTime":"2025-12-03T06:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.736936 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.736963 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.736946 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:41 crc kubenswrapper[4818]: E1203 06:28:41.737053 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:41 crc kubenswrapper[4818]: E1203 06:28:41.737147 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:41 crc kubenswrapper[4818]: E1203 06:28:41.737244 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.808696 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.808736 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.808746 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.808760 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.808774 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:41Z","lastTransitionTime":"2025-12-03T06:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.912336 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.912380 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.912394 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.912411 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:41 crc kubenswrapper[4818]: I1203 06:28:41.912422 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:41Z","lastTransitionTime":"2025-12-03T06:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.018546 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.018585 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.018594 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.018865 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.018903 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:42Z","lastTransitionTime":"2025-12-03T06:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.121441 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.121477 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.121488 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.121501 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.121510 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:42Z","lastTransitionTime":"2025-12-03T06:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.223152 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.223216 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.223229 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.223245 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.223258 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:42Z","lastTransitionTime":"2025-12-03T06:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.326609 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.326658 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.326670 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.326687 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.326700 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:42Z","lastTransitionTime":"2025-12-03T06:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.430329 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.430402 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.430414 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.430429 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.430439 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:42Z","lastTransitionTime":"2025-12-03T06:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.533210 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.533249 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.533260 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.533275 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.533287 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:42Z","lastTransitionTime":"2025-12-03T06:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.636350 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.636411 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.636432 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.636458 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.636478 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:42Z","lastTransitionTime":"2025-12-03T06:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.737031 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:42 crc kubenswrapper[4818]: E1203 06:28:42.737155 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.738669 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.738887 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.739083 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.739229 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.739355 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:42Z","lastTransitionTime":"2025-12-03T06:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.755103 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.771211 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.791879 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.822232 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:28:17Z\\\",\\\"message\\\":\\\"37375 6453 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:28:17.137393 6453 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:28:17.138535 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:28:17.138567 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:28:17.138575 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:28:17.138607 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:28:17.138627 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:28:17.138664 6453 factory.go:656] Stopping watch factory\\\\nI1203 06:28:17.138688 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:28:17.138729 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:28:17.138743 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:28:17.138751 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:28:17.138758 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:28:17.138767 6453 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:28:17.138781 6453 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:28:17.138948 6453 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:28:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.839689 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d8af736-67f3-4db0-a7a5-a24af4222b86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:28:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gq8r2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.841382 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.841419 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.841428 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.841441 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.841450 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:42Z","lastTransitionTime":"2025-12-03T06:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.859130 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.873497 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.895462 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.910691 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.925699 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.939792 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.943767 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.943863 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.943891 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.943918 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.943941 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:42Z","lastTransitionTime":"2025-12-03T06:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.956961 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b92f0662c82eef1fb0023107da93fead648febd6e3880b77ad264ced57de7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:28:32Z\\\",\\\"message\\\":\\\"2025-12-03T06:27:47+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a1c583e8-fae6-4b94-af4c-e7b26ade3559\\\\n2025-12-03T06:27:47+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a1c583e8-fae6-4b94-af4c-e7b26ade3559 to /host/opt/cni/bin/\\\\n2025-12-03T06:27:47Z [verbose] multus-daemon started\\\\n2025-12-03T06:27:47Z [verbose] Readiness Indicator file check\\\\n2025-12-03T06:28:32Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.969614 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d407a21-4cfd-4505-865b-55e30c39b4e8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8047009041ecc11795cad4ced8b06bc2a4810c42023810db409854dfcaae2275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5faa1cad79099f32db2b2895d0fb547031ee6dfa6c4a9e306f355b93e4c7c959\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a192429ab94e49222181fd50bb4e9597966a1b6adb38c8ebaf0eb31fcf6b8bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.983901 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:42 crc kubenswrapper[4818]: I1203 06:28:42.997528 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:42Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.011759 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.027655 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.043039 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://750933f24a9ca216d9efc07f0840656da96538ddfddf03931295e18ffbe79778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25d96ab68b06a691db5cdfc715ad128e22902ba1d747a1b468be6283ae3579cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:43Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.046569 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.046600 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.046609 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.046621 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.046632 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:43Z","lastTransitionTime":"2025-12-03T06:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.149586 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.149674 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.149703 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.149735 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.149755 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:43Z","lastTransitionTime":"2025-12-03T06:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.255012 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.255081 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.255101 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.255124 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.255141 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:43Z","lastTransitionTime":"2025-12-03T06:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.358420 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.358496 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.358521 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.358551 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.358572 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:43Z","lastTransitionTime":"2025-12-03T06:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.463305 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.463348 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.463359 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.463374 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.463385 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:43Z","lastTransitionTime":"2025-12-03T06:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.565789 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.565851 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.565867 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.565882 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.565894 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:43Z","lastTransitionTime":"2025-12-03T06:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.668919 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.669007 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.669021 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.669036 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.669046 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:43Z","lastTransitionTime":"2025-12-03T06:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.736966 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:43 crc kubenswrapper[4818]: E1203 06:28:43.737101 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.737297 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.737413 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:43 crc kubenswrapper[4818]: E1203 06:28:43.737556 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:43 crc kubenswrapper[4818]: E1203 06:28:43.737934 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.771875 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.771923 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.771941 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.771963 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.771981 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:43Z","lastTransitionTime":"2025-12-03T06:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.875576 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.875646 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.875666 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.875694 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.875718 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:43Z","lastTransitionTime":"2025-12-03T06:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.978610 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.978672 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.978691 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.978714 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:43 crc kubenswrapper[4818]: I1203 06:28:43.978731 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:43Z","lastTransitionTime":"2025-12-03T06:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.081776 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.081839 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.081854 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.081868 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.081879 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:44Z","lastTransitionTime":"2025-12-03T06:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.183699 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.183740 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.183754 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.183777 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.183792 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:44Z","lastTransitionTime":"2025-12-03T06:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.286472 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.286507 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.286545 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.286560 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.286570 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:44Z","lastTransitionTime":"2025-12-03T06:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.389995 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.390060 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.390079 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.390102 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.390118 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:44Z","lastTransitionTime":"2025-12-03T06:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.493042 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.493101 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.493119 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.493143 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.493160 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:44Z","lastTransitionTime":"2025-12-03T06:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.493186 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:28:44 crc kubenswrapper[4818]: E1203 06:28:44.493373 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:29:48.49335271 +0000 UTC m=+146.184961482 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.594793 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:44 crc kubenswrapper[4818]: E1203 06:28:44.595080 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.595061 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:44 crc kubenswrapper[4818]: E1203 06:28:44.595180 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:29:48.595150195 +0000 UTC m=+146.286758987 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.595225 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.595309 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:44 crc kubenswrapper[4818]: E1203 06:28:44.595419 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:28:44 crc kubenswrapper[4818]: E1203 06:28:44.595446 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:28:44 crc kubenswrapper[4818]: E1203 06:28:44.595473 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:28:44 crc kubenswrapper[4818]: E1203 06:28:44.595491 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:28:44 crc kubenswrapper[4818]: E1203 06:28:44.595513 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:29:48.595489664 +0000 UTC m=+146.287098456 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:28:44 crc kubenswrapper[4818]: E1203 06:28:44.595544 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:29:48.595525315 +0000 UTC m=+146.287134107 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:28:44 crc kubenswrapper[4818]: E1203 06:28:44.595614 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:28:44 crc kubenswrapper[4818]: E1203 06:28:44.595641 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:28:44 crc kubenswrapper[4818]: E1203 06:28:44.595663 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:28:44 crc kubenswrapper[4818]: E1203 06:28:44.595738 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:29:48.595715389 +0000 UTC m=+146.287324181 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.596135 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.596193 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.596210 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.596233 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.596250 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:44Z","lastTransitionTime":"2025-12-03T06:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.698990 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.699076 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.699092 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.699173 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.699191 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:44Z","lastTransitionTime":"2025-12-03T06:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.737633 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:44 crc kubenswrapper[4818]: E1203 06:28:44.737799 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.802186 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.802239 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.802255 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.802278 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.802294 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:44Z","lastTransitionTime":"2025-12-03T06:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.905436 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.905528 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.905551 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.905582 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:44 crc kubenswrapper[4818]: I1203 06:28:44.905625 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:44Z","lastTransitionTime":"2025-12-03T06:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.008528 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.008581 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.008595 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.008612 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.008624 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:45Z","lastTransitionTime":"2025-12-03T06:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.110697 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.110733 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.110744 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.110759 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.110769 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:45Z","lastTransitionTime":"2025-12-03T06:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.211593 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.211781 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.211876 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.211912 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.211983 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:45Z","lastTransitionTime":"2025-12-03T06:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:45 crc kubenswrapper[4818]: E1203 06:28:45.234726 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.239371 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.239416 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.239428 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.239444 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.239457 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:45Z","lastTransitionTime":"2025-12-03T06:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:45 crc kubenswrapper[4818]: E1203 06:28:45.254122 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.257902 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.257930 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.257939 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.257952 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.257962 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:45Z","lastTransitionTime":"2025-12-03T06:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:45 crc kubenswrapper[4818]: E1203 06:28:45.270989 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.274605 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.274634 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.274641 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.274651 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.274659 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:45Z","lastTransitionTime":"2025-12-03T06:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:45 crc kubenswrapper[4818]: E1203 06:28:45.289486 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.293649 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.293675 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.293683 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.293695 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.293704 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:45Z","lastTransitionTime":"2025-12-03T06:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:45 crc kubenswrapper[4818]: E1203 06:28:45.305851 4818 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b45b9166-fc05-4a5d-ae82-e08efaed5d6b\\\",\\\"systemUUID\\\":\\\"af2f867d-be4e-4cab-a642-bf5ce3ee5141\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:45Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:45 crc kubenswrapper[4818]: E1203 06:28:45.306003 4818 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.307895 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.307929 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.307940 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.307957 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.307969 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:45Z","lastTransitionTime":"2025-12-03T06:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.410056 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.410096 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.410105 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.410120 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.410131 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:45Z","lastTransitionTime":"2025-12-03T06:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.512646 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.512683 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.512692 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.512705 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.512714 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:45Z","lastTransitionTime":"2025-12-03T06:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.615836 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.615940 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.615959 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.615984 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.615997 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:45Z","lastTransitionTime":"2025-12-03T06:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.718480 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.718538 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.718550 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.718565 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.718576 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:45Z","lastTransitionTime":"2025-12-03T06:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.737150 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.737231 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:45 crc kubenswrapper[4818]: E1203 06:28:45.737366 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.737231 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:45 crc kubenswrapper[4818]: E1203 06:28:45.737432 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:45 crc kubenswrapper[4818]: E1203 06:28:45.737567 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.820415 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.820450 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.820459 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.820473 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.820484 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:45Z","lastTransitionTime":"2025-12-03T06:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.924607 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.924652 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.924660 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.924677 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:45 crc kubenswrapper[4818]: I1203 06:28:45.924691 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:45Z","lastTransitionTime":"2025-12-03T06:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.027391 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.027436 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.027446 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.027460 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.027471 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:46Z","lastTransitionTime":"2025-12-03T06:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.130627 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.130682 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.130695 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.130711 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.130723 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:46Z","lastTransitionTime":"2025-12-03T06:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.233495 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.233551 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.233568 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.233590 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.233608 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:46Z","lastTransitionTime":"2025-12-03T06:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.336660 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.336705 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.336718 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.336736 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.336750 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:46Z","lastTransitionTime":"2025-12-03T06:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.439131 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.439172 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.439183 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.439198 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.439208 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:46Z","lastTransitionTime":"2025-12-03T06:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.542185 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.542516 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.542683 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.542873 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.543024 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:46Z","lastTransitionTime":"2025-12-03T06:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.646523 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.646562 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.646572 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.646586 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.646597 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:46Z","lastTransitionTime":"2025-12-03T06:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.737397 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:46 crc kubenswrapper[4818]: E1203 06:28:46.737581 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.749355 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.749388 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.749400 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.749414 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.749424 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:46Z","lastTransitionTime":"2025-12-03T06:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.851845 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.851890 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.851906 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.851931 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.851949 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:46Z","lastTransitionTime":"2025-12-03T06:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.954451 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.954485 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.954494 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.954508 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:46 crc kubenswrapper[4818]: I1203 06:28:46.954518 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:46Z","lastTransitionTime":"2025-12-03T06:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.057161 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.057206 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.057215 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.057229 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.057241 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:47Z","lastTransitionTime":"2025-12-03T06:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.161012 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.161067 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.161080 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.161100 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.161118 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:47Z","lastTransitionTime":"2025-12-03T06:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.263973 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.264195 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.264314 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.264479 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.264607 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:47Z","lastTransitionTime":"2025-12-03T06:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.368336 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.368379 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.368391 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.368406 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.368418 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:47Z","lastTransitionTime":"2025-12-03T06:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.471107 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.471166 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.471219 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.471249 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.471269 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:47Z","lastTransitionTime":"2025-12-03T06:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.574672 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.575060 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.575439 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.575725 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.575977 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:47Z","lastTransitionTime":"2025-12-03T06:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.678849 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.678896 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.678906 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.678920 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.678931 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:47Z","lastTransitionTime":"2025-12-03T06:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.736616 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:47 crc kubenswrapper[4818]: E1203 06:28:47.737087 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.738130 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.738505 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:47 crc kubenswrapper[4818]: E1203 06:28:47.739236 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:47 crc kubenswrapper[4818]: E1203 06:28:47.739358 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.739888 4818 scope.go:117] "RemoveContainer" containerID="fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.783073 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.783520 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.783533 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.783549 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.783562 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:47Z","lastTransitionTime":"2025-12-03T06:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.887060 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.887108 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.887153 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.887177 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.887194 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:47Z","lastTransitionTime":"2025-12-03T06:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.989065 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.989096 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.989104 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.989131 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:47 crc kubenswrapper[4818]: I1203 06:28:47.989141 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:47Z","lastTransitionTime":"2025-12-03T06:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.092039 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.092080 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.092091 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.092106 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.092119 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:48Z","lastTransitionTime":"2025-12-03T06:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.183578 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovnkube-controller/2.log" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.186292 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerStarted","Data":"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024"} Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.186759 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.194676 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.194713 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.194723 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.194740 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.194751 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:48Z","lastTransitionTime":"2025-12-03T06:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.214236 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.229290 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.245333 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.255606 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.267120 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b92f0662c82eef1fb0023107da93fead648febd6e3880b77ad264ced57de7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:28:32Z\\\",\\\"message\\\":\\\"2025-12-03T06:27:47+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a1c583e8-fae6-4b94-af4c-e7b26ade3559\\\\n2025-12-03T06:27:47+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a1c583e8-fae6-4b94-af4c-e7b26ade3559 to /host/opt/cni/bin/\\\\n2025-12-03T06:27:47Z [verbose] multus-daemon started\\\\n2025-12-03T06:27:47Z [verbose] Readiness Indicator file check\\\\n2025-12-03T06:28:32Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.281333 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d407a21-4cfd-4505-865b-55e30c39b4e8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8047009041ecc11795cad4ced8b06bc2a4810c42023810db409854dfcaae2275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5faa1cad79099f32db2b2895d0fb547031ee6dfa6c4a9e306f355b93e4c7c959\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a192429ab94e49222181fd50bb4e9597966a1b6adb38c8ebaf0eb31fcf6b8bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.296685 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.296739 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.296748 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.296761 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.296770 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:48Z","lastTransitionTime":"2025-12-03T06:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.297666 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.311915 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.330402 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.340356 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.358485 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://750933f24a9ca216d9efc07f0840656da96538ddfddf03931295e18ffbe79778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25d96ab68b06a691db5cdfc715ad128e22902ba1d747a1b468be6283ae3579cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.383684 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.398160 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.399102 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.399149 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.399159 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.399171 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.399182 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:48Z","lastTransitionTime":"2025-12-03T06:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.415280 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.434570 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:28:17Z\\\",\\\"message\\\":\\\"37375 6453 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:28:17.137393 6453 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:28:17.138535 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:28:17.138567 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:28:17.138575 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:28:17.138607 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:28:17.138627 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:28:17.138664 6453 factory.go:656] Stopping watch factory\\\\nI1203 06:28:17.138688 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:28:17.138729 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:28:17.138743 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:28:17.138751 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:28:17.138758 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:28:17.138767 6453 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:28:17.138781 6453 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:28:17.138948 6453 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:28:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.444489 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d8af736-67f3-4db0-a7a5-a24af4222b86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:28:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gq8r2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.458269 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.469577 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:48Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.501979 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.502042 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.502056 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.502080 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.502107 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:48Z","lastTransitionTime":"2025-12-03T06:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.605468 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.605507 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.605516 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.605528 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.605539 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:48Z","lastTransitionTime":"2025-12-03T06:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.708491 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.708540 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.708552 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.708575 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.708588 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:48Z","lastTransitionTime":"2025-12-03T06:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.737987 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:48 crc kubenswrapper[4818]: E1203 06:28:48.738177 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.811835 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.811882 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.811895 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.811912 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.811925 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:48Z","lastTransitionTime":"2025-12-03T06:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.914871 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.914934 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.914948 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.914969 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:48 crc kubenswrapper[4818]: I1203 06:28:48.914992 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:48Z","lastTransitionTime":"2025-12-03T06:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.017404 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.017459 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.017469 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.017487 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.017503 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:49Z","lastTransitionTime":"2025-12-03T06:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.120416 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.120475 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.120493 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.120516 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.120536 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:49Z","lastTransitionTime":"2025-12-03T06:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.192019 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovnkube-controller/3.log" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.193108 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovnkube-controller/2.log" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.196134 4818 generic.go:334] "Generic (PLEG): container finished" podID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerID="0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024" exitCode=1 Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.196176 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerDied","Data":"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024"} Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.196240 4818 scope.go:117] "RemoveContainer" containerID="fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.197255 4818 scope.go:117] "RemoveContainer" containerID="0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024" Dec 03 06:28:49 crc kubenswrapper[4818]: E1203 06:28:49.197519 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.219255 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.224479 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.224520 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.224535 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.224558 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.224574 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:49Z","lastTransitionTime":"2025-12-03T06:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.237630 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.253942 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.281062 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb1669665e8a5bdecf20e7d815dc871a15318442e20a55a4c436e7c727223bff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:28:17Z\\\",\\\"message\\\":\\\"37375 6453 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1203 06:28:17.137393 6453 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 06:28:17.138535 6453 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 06:28:17.138567 6453 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 06:28:17.138575 6453 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 06:28:17.138607 6453 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 06:28:17.138627 6453 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 06:28:17.138664 6453 factory.go:656] Stopping watch factory\\\\nI1203 06:28:17.138688 6453 ovnkube.go:599] Stopped ovnkube\\\\nI1203 06:28:17.138729 6453 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 06:28:17.138743 6453 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 06:28:17.138751 6453 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 06:28:17.138758 6453 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 06:28:17.138767 6453 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 06:28:17.138781 6453 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 06:28:17.138948 6453 ovnkube.go:137] failed to run ovnkube: [failed to start network contr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:28:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:28:48Z\\\",\\\"message\\\":\\\"Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.110\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1203 06:28:48.602969 6868 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node networ\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:28:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.295966 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d8af736-67f3-4db0-a7a5-a24af4222b86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:28:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gq8r2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.312962 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.327598 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.328330 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.328458 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.328547 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.328644 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.328725 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:49Z","lastTransitionTime":"2025-12-03T06:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.348122 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.361857 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.373875 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.382979 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.394932 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b92f0662c82eef1fb0023107da93fead648febd6e3880b77ad264ced57de7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:28:32Z\\\",\\\"message\\\":\\\"2025-12-03T06:27:47+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a1c583e8-fae6-4b94-af4c-e7b26ade3559\\\\n2025-12-03T06:27:47+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a1c583e8-fae6-4b94-af4c-e7b26ade3559 to /host/opt/cni/bin/\\\\n2025-12-03T06:27:47Z [verbose] multus-daemon started\\\\n2025-12-03T06:27:47Z [verbose] Readiness Indicator file check\\\\n2025-12-03T06:28:32Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.405385 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d407a21-4cfd-4505-865b-55e30c39b4e8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8047009041ecc11795cad4ced8b06bc2a4810c42023810db409854dfcaae2275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5faa1cad79099f32db2b2895d0fb547031ee6dfa6c4a9e306f355b93e4c7c959\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a192429ab94e49222181fd50bb4e9597966a1b6adb38c8ebaf0eb31fcf6b8bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.417934 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.429764 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.430763 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.430808 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.430835 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.430850 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.430860 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:49Z","lastTransitionTime":"2025-12-03T06:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.443646 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.453086 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.466608 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://750933f24a9ca216d9efc07f0840656da96538ddfddf03931295e18ffbe79778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25d96ab68b06a691db5cdfc715ad128e22902ba1d747a1b468be6283ae3579cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:49Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.533776 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.534191 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.534408 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.534595 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.534774 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:49Z","lastTransitionTime":"2025-12-03T06:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.638488 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.638914 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.638999 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.639070 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.639144 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:49Z","lastTransitionTime":"2025-12-03T06:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.737524 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:49 crc kubenswrapper[4818]: E1203 06:28:49.737733 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.737536 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:49 crc kubenswrapper[4818]: E1203 06:28:49.737949 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.737524 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:49 crc kubenswrapper[4818]: E1203 06:28:49.738083 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.741388 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.741446 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.741465 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.741489 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.741508 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:49Z","lastTransitionTime":"2025-12-03T06:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.844462 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.844535 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.844559 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.844589 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.844613 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:49Z","lastTransitionTime":"2025-12-03T06:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.947994 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.948037 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.948048 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.948066 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:49 crc kubenswrapper[4818]: I1203 06:28:49.948081 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:49Z","lastTransitionTime":"2025-12-03T06:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.050364 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.050443 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.050467 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.050495 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.050519 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:50Z","lastTransitionTime":"2025-12-03T06:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.153798 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.153941 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.153960 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.153986 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.154013 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:50Z","lastTransitionTime":"2025-12-03T06:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.202454 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovnkube-controller/3.log" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.207296 4818 scope.go:117] "RemoveContainer" containerID="0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024" Dec 03 06:28:50 crc kubenswrapper[4818]: E1203 06:28:50.207515 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.224753 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10740ef3-c8fb-4663-86ef-b46c1f969ffb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e7c94e82115bfa77d18118440fccdb28d5410bcbd2b8b5c9753a43a2597fc16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3c538d401f0751a1b4f3afd867dc61835c65b5343c8bd6fc33fc74b24ebd108\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca4e8f3d97d86d84f59ed5cf84e278f5d0aa4a5de36d0a7bea8911a8c702eb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afa62121930b46da5d978ecdd9f2c83489d215b8f2a1028b5e05f504add1298d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2db53f3589f7b2f0e9a3f50ce1bce6667f7ae9a4eebf9c391ab966c6bd62189\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cd7460f9bb70beaf526feddef7c4e701db7e207a9a1552bc42f9de3f826cafc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2cca20e8ade8fd061136965a6258ad958db547020c35a30d26a75184f74bb3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qxnz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pjpw7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.238708 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd398f0d-786f-44e7-a64b-7cf7516d5d54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39220db5e9ff94a72ca86fce985a3e5169bf0e52991575d03dfa22d7e397c3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-trn6s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t9fzk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.260073 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.260154 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.260166 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.260183 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.260195 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:50Z","lastTransitionTime":"2025-12-03T06:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.264058 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77089997-b1c0-42f0-84a5-66576a2d92f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c20a119a3020cd3372ac8425726674dbd526dfb1de00394e76550f3c2c77cda4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f34dd10105b5ee0c26c8e998a1a98fa80b1430b79d1495981f9012d9581e3ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://867d6a1a608220b12f11751cfcc32ff34fc7edd5f4698aa9a9ce07f7fb825a86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ccb5ee8d656cf3daa2119fdacf733bbe0539f04d9e31aa11246a45c3efbf2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345a8d75a4fec51e2cbf09b302d53fe95225e181b926f1b604176b309674dc44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a0eb49ba82cc23a73af588b040cd6fad1c089ce2b6aa4e27b417a765e75ba72c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f62e29587d18f7330f3420c60290396b017d265c55cf5597721554ef533c2a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://23ab976646322b1bf1cc4540b3dde1641af13e5621ce1aa88afd77ea565fce1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.284141 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"558ef6b2-cb0b-4631-b025-8fe23b244a88\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 06:27:35.044211 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 06:27:35.045628 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-337011669/tls.crt::/tmp/serving-cert-337011669/tls.key\\\\\\\"\\\\nI1203 06:27:40.530255 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 06:27:40.541072 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 06:27:40.541101 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 06:27:40.541139 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 06:27:40.541145 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 06:27:40.546606 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 06:27:40.546634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 06:27:40.546636 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 06:27:40.546640 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 06:27:40.546655 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 06:27:40.546659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 06:27:40.546662 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 06:27:40.546665 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1203 06:27:40.547836 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.300954 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c703cd7346b57d61dba108a247d2f4c870d2d824e9a20f1225afebc7315cd14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.312406 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-m4v69" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1b5b453f-6556-4a63-ace6-494d845ff04b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6da86f2035ed1278489389ec7c06d1b08b5c2722e3df8c688401f35c17e1653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pf6mj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-m4v69\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.327491 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ctpzs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b25b836-b639-4111-bcea-af7cc791ea32\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b92f0662c82eef1fb0023107da93fead648febd6e3880b77ad264ced57de7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:28:32Z\\\",\\\"message\\\":\\\"2025-12-03T06:27:47+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a1c583e8-fae6-4b94-af4c-e7b26ade3559\\\\n2025-12-03T06:27:47+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a1c583e8-fae6-4b94-af4c-e7b26ade3559 to /host/opt/cni/bin/\\\\n2025-12-03T06:27:47Z [verbose] multus-daemon started\\\\n2025-12-03T06:27:47Z [verbose] Readiness Indicator file check\\\\n2025-12-03T06:28:32Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ctpzs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.340144 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57192c6e-16e0-458f-93aa-9af8af4eea19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://750933f24a9ca216d9efc07f0840656da96538ddfddf03931295e18ffbe79778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25d96ab68b06a691db5cdfc715ad128e22902ba1d747a1b468be6283ae3579cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2blz4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:58Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vr6bc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.352124 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d407a21-4cfd-4505-865b-55e30c39b4e8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8047009041ecc11795cad4ced8b06bc2a4810c42023810db409854dfcaae2275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5faa1cad79099f32db2b2895d0fb547031ee6dfa6c4a9e306f355b93e4c7c959\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a192429ab94e49222181fd50bb4e9597966a1b6adb38c8ebaf0eb31fcf6b8bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.362523 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.362581 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.362597 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.362616 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.362638 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:50Z","lastTransitionTime":"2025-12-03T06:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.365842 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.378125 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.391162 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bc6fffa8335c24a4168c74867014491cbe5bcfda213ff8e1a0c62a8ec89cfc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90c961008baeaea93c4df9cdbca6150a39e90e452c8d9d750642ee9e03309aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.402117 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2ds8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84d9588b-f409-4fc0-9f8e-3ec975189e5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://981e6d146b299fb35c27bf80b29bce1679288308b6c761799289c4cfc0371e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xq84n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2ds8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.413312 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d8af736-67f3-4db0-a7a5-a24af4222b86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fgxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:28:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-gq8r2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.430271 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50064a3ca3df425d1baf68bde5fdd3db37a3e8432364498a23d48e845999057b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.441689 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.453389 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.464947 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.464989 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.464999 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.465015 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.465025 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:50Z","lastTransitionTime":"2025-12-03T06:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.475135 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1e30fe6-0b94-4c95-b941-c0623729e123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T06:28:48Z\\\",\\\"message\\\":\\\"Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.110\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1203 06:28:48.602969 6868 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node networ\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T06:28:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrtqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ktm8k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:50Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.567068 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.567114 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.567131 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.567151 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.567165 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:50Z","lastTransitionTime":"2025-12-03T06:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.669526 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.669963 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.670138 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.670314 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.670655 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:50Z","lastTransitionTime":"2025-12-03T06:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.737101 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:50 crc kubenswrapper[4818]: E1203 06:28:50.737473 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.747027 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.773561 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.773613 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.773625 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.773643 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.773655 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:50Z","lastTransitionTime":"2025-12-03T06:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.876145 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.876185 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.876199 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.876219 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.876235 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:50Z","lastTransitionTime":"2025-12-03T06:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.978375 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.978442 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.978457 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.978474 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:50 crc kubenswrapper[4818]: I1203 06:28:50.978486 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:50Z","lastTransitionTime":"2025-12-03T06:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.081415 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.081485 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.081507 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.081537 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.081557 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:51Z","lastTransitionTime":"2025-12-03T06:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.183332 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.183376 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.183388 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.183403 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.183413 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:51Z","lastTransitionTime":"2025-12-03T06:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.286859 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.286925 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.286938 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.286958 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.286972 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:51Z","lastTransitionTime":"2025-12-03T06:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.394481 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.394573 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.394682 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.394785 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.395695 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:51Z","lastTransitionTime":"2025-12-03T06:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.498599 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.498664 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.498733 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.498762 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.498784 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:51Z","lastTransitionTime":"2025-12-03T06:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.601901 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.601961 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.601989 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.602003 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.602012 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:51Z","lastTransitionTime":"2025-12-03T06:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.704255 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.704316 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.704333 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.704357 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.704374 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:51Z","lastTransitionTime":"2025-12-03T06:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.737061 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.737106 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.737107 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:51 crc kubenswrapper[4818]: E1203 06:28:51.737252 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:51 crc kubenswrapper[4818]: E1203 06:28:51.737484 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:51 crc kubenswrapper[4818]: E1203 06:28:51.737622 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.807523 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.807593 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.807608 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.807629 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.807656 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:51Z","lastTransitionTime":"2025-12-03T06:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.909625 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.909662 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.909671 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.909685 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:51 crc kubenswrapper[4818]: I1203 06:28:51.909695 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:51Z","lastTransitionTime":"2025-12-03T06:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.012802 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.012897 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.012920 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.012952 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.012973 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:52Z","lastTransitionTime":"2025-12-03T06:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.115440 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.115471 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.115482 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.115497 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.115508 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:52Z","lastTransitionTime":"2025-12-03T06:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.218098 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.218170 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.218203 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.218219 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.218228 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:52Z","lastTransitionTime":"2025-12-03T06:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.320441 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.320508 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.320518 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.320537 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.320550 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:52Z","lastTransitionTime":"2025-12-03T06:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.423483 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.423561 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.423621 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.423657 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.423681 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:52Z","lastTransitionTime":"2025-12-03T06:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.526652 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.526695 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.526707 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.526723 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.526733 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:52Z","lastTransitionTime":"2025-12-03T06:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.628873 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.628927 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.628940 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.628961 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.628973 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:52Z","lastTransitionTime":"2025-12-03T06:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.731290 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.731369 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.731381 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.731396 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.731408 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:52Z","lastTransitionTime":"2025-12-03T06:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.737160 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:52 crc kubenswrapper[4818]: E1203 06:28:52.737341 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.753699 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d407a21-4cfd-4505-865b-55e30c39b4e8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:28:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8047009041ecc11795cad4ced8b06bc2a4810c42023810db409854dfcaae2275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5faa1cad79099f32db2b2895d0fb547031ee6dfa6c4a9e306f355b93e4c7c959\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a192429ab94e49222181fd50bb4e9597966a1b6adb38c8ebaf0eb31fcf6b8bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://471dab777fc867f33dad58818c1b15c31207ce87d09e9b6241a50391ef70c3aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T06:27:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.768445 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b62664b-fb5e-44d5-b6b3-efa48b2c5dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c742779d9c36c2b43bf96284a70dce35e39236e942b4acb5cd7561959fb01783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3bf5e658488569681b71517604c9be4ab2e9c2ba8b460f14de134b5614109f7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d7542ffeff67f18282a211fbb5475927d9bc58bf3755515e85f1b4e92a1a079\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T06:27:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T06:27:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.781859 4818 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T06:27:40Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T06:28:52Z is after 2025-08-24T17:21:41Z" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.822988 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-2ds8v" podStartSLOduration=67.82294833 podStartE2EDuration="1m7.82294833s" podCreationTimestamp="2025-12-03 06:27:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:28:52.8229352 +0000 UTC m=+90.514543952" watchObservedRunningTime="2025-12-03 06:28:52.82294833 +0000 UTC m=+90.514557092" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.833468 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.833505 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.833516 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.833533 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.833545 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:52Z","lastTransitionTime":"2025-12-03T06:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.858237 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vr6bc" podStartSLOduration=66.858207848 podStartE2EDuration="1m6.858207848s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:28:52.844884057 +0000 UTC m=+90.536492829" watchObservedRunningTime="2025-12-03 06:28:52.858207848 +0000 UTC m=+90.549816610" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.876765 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=2.87674485 podStartE2EDuration="2.87674485s" podCreationTimestamp="2025-12-03 06:28:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:28:52.858786773 +0000 UTC m=+90.550395535" watchObservedRunningTime="2025-12-03 06:28:52.87674485 +0000 UTC m=+90.568353602" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.937034 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.937105 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.937118 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.937145 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.937158 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:52Z","lastTransitionTime":"2025-12-03T06:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.966397 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-pjpw7" podStartSLOduration=66.966369552 podStartE2EDuration="1m6.966369552s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:28:52.965859349 +0000 UTC m=+90.657468101" watchObservedRunningTime="2025-12-03 06:28:52.966369552 +0000 UTC m=+90.657978304" Dec 03 06:28:52 crc kubenswrapper[4818]: I1203 06:28:52.981781 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podStartSLOduration=66.981759075 podStartE2EDuration="1m6.981759075s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:28:52.981213072 +0000 UTC m=+90.672821834" watchObservedRunningTime="2025-12-03 06:28:52.981759075 +0000 UTC m=+90.673367827" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.012331 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=70.012309476 podStartE2EDuration="1m10.012309476s" podCreationTimestamp="2025-12-03 06:27:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:28:53.00967089 +0000 UTC m=+90.701279652" watchObservedRunningTime="2025-12-03 06:28:53.012309476 +0000 UTC m=+90.703918228" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.026551 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=73.02653454 podStartE2EDuration="1m13.02653454s" podCreationTimestamp="2025-12-03 06:27:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:28:53.025764411 +0000 UTC m=+90.717373163" watchObservedRunningTime="2025-12-03 06:28:53.02653454 +0000 UTC m=+90.718143292" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.040312 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.040359 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.040368 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.040386 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.040399 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:53Z","lastTransitionTime":"2025-12-03T06:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.071364 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-m4v69" podStartSLOduration=68.071340676 podStartE2EDuration="1m8.071340676s" podCreationTimestamp="2025-12-03 06:27:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:28:53.057161603 +0000 UTC m=+90.748770355" watchObservedRunningTime="2025-12-03 06:28:53.071340676 +0000 UTC m=+90.762949428" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.143221 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.143270 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.143280 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.143495 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.143510 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:53Z","lastTransitionTime":"2025-12-03T06:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.247063 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.247129 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.247138 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.247162 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.247191 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:53Z","lastTransitionTime":"2025-12-03T06:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.351024 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.351089 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.351111 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.351141 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.351164 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:53Z","lastTransitionTime":"2025-12-03T06:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.454110 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.454164 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.454175 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.454192 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.454204 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:53Z","lastTransitionTime":"2025-12-03T06:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.556757 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.556810 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.556846 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.556865 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.556877 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:53Z","lastTransitionTime":"2025-12-03T06:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.659480 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.659536 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.659546 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.659565 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.659578 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:53Z","lastTransitionTime":"2025-12-03T06:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.736981 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.737071 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.737145 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:53 crc kubenswrapper[4818]: E1203 06:28:53.737233 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:53 crc kubenswrapper[4818]: E1203 06:28:53.737312 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:53 crc kubenswrapper[4818]: E1203 06:28:53.737396 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.762360 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.762428 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.762448 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.762472 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.762490 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:53Z","lastTransitionTime":"2025-12-03T06:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.865936 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.866037 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.866055 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.866084 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.866104 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:53Z","lastTransitionTime":"2025-12-03T06:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.968935 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.968988 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.969000 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.969022 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:53 crc kubenswrapper[4818]: I1203 06:28:53.969037 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:53Z","lastTransitionTime":"2025-12-03T06:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.076431 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.076498 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.076511 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.076532 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.076547 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:54Z","lastTransitionTime":"2025-12-03T06:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.180631 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.180703 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.180723 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.180752 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.180773 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:54Z","lastTransitionTime":"2025-12-03T06:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.284354 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.284406 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.284416 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.284435 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.284486 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:54Z","lastTransitionTime":"2025-12-03T06:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.387572 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.387640 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.387653 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.387694 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.387707 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:54Z","lastTransitionTime":"2025-12-03T06:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.492019 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.492169 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.492190 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.492224 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.492344 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:54Z","lastTransitionTime":"2025-12-03T06:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.596474 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.596561 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.596574 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.596594 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.596611 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:54Z","lastTransitionTime":"2025-12-03T06:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.701162 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.701248 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.701272 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.701304 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.701328 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:54Z","lastTransitionTime":"2025-12-03T06:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.737317 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:54 crc kubenswrapper[4818]: E1203 06:28:54.737467 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.804916 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.804986 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.805008 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.805080 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.805106 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:54Z","lastTransitionTime":"2025-12-03T06:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.908151 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.908218 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.908235 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.908259 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:54 crc kubenswrapper[4818]: I1203 06:28:54.908274 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:54Z","lastTransitionTime":"2025-12-03T06:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.010525 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.010563 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.010573 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.010586 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.010596 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:55Z","lastTransitionTime":"2025-12-03T06:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.113511 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.113560 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.113573 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.113595 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.113610 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:55Z","lastTransitionTime":"2025-12-03T06:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.217122 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.217187 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.217199 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.217220 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.217233 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:55Z","lastTransitionTime":"2025-12-03T06:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.320482 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.320515 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.320527 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.320543 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.320554 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:55Z","lastTransitionTime":"2025-12-03T06:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.424428 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.424504 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.424529 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.424566 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.424592 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:55Z","lastTransitionTime":"2025-12-03T06:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.527378 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.527452 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.527479 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.527506 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.527528 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:55Z","lastTransitionTime":"2025-12-03T06:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.566735 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.566782 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.566793 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.566811 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.566840 4818 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T06:28:55Z","lastTransitionTime":"2025-12-03T06:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.628246 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-ctpzs" podStartSLOduration=69.62822166 podStartE2EDuration="1m9.62822166s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:28:53.072078875 +0000 UTC m=+90.763687627" watchObservedRunningTime="2025-12-03 06:28:55.62822166 +0000 UTC m=+93.319830442" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.629109 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz"] Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.629615 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.632410 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.632518 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.632666 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.633776 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.649136 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=70.64910428 podStartE2EDuration="1m10.64910428s" podCreationTimestamp="2025-12-03 06:27:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:28:55.648132695 +0000 UTC m=+93.339741447" watchObservedRunningTime="2025-12-03 06:28:55.64910428 +0000 UTC m=+93.340713052" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.727922 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e2fd795f-95bc-401e-92f0-30e1d99130bc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lgqfz\" (UID: \"e2fd795f-95bc-401e-92f0-30e1d99130bc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.728258 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e2fd795f-95bc-401e-92f0-30e1d99130bc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lgqfz\" (UID: \"e2fd795f-95bc-401e-92f0-30e1d99130bc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.728432 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e2fd795f-95bc-401e-92f0-30e1d99130bc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lgqfz\" (UID: \"e2fd795f-95bc-401e-92f0-30e1d99130bc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.728576 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2fd795f-95bc-401e-92f0-30e1d99130bc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lgqfz\" (UID: \"e2fd795f-95bc-401e-92f0-30e1d99130bc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.728734 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e2fd795f-95bc-401e-92f0-30e1d99130bc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lgqfz\" (UID: \"e2fd795f-95bc-401e-92f0-30e1d99130bc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.737357 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:55 crc kubenswrapper[4818]: E1203 06:28:55.737582 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.737400 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:55 crc kubenswrapper[4818]: E1203 06:28:55.738268 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.738637 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:55 crc kubenswrapper[4818]: E1203 06:28:55.738890 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.830174 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e2fd795f-95bc-401e-92f0-30e1d99130bc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lgqfz\" (UID: \"e2fd795f-95bc-401e-92f0-30e1d99130bc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.830424 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e2fd795f-95bc-401e-92f0-30e1d99130bc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lgqfz\" (UID: \"e2fd795f-95bc-401e-92f0-30e1d99130bc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.830669 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e2fd795f-95bc-401e-92f0-30e1d99130bc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lgqfz\" (UID: \"e2fd795f-95bc-401e-92f0-30e1d99130bc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.830785 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e2fd795f-95bc-401e-92f0-30e1d99130bc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lgqfz\" (UID: \"e2fd795f-95bc-401e-92f0-30e1d99130bc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.831262 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2fd795f-95bc-401e-92f0-30e1d99130bc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lgqfz\" (UID: \"e2fd795f-95bc-401e-92f0-30e1d99130bc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.832164 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e2fd795f-95bc-401e-92f0-30e1d99130bc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lgqfz\" (UID: \"e2fd795f-95bc-401e-92f0-30e1d99130bc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.832067 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e2fd795f-95bc-401e-92f0-30e1d99130bc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lgqfz\" (UID: \"e2fd795f-95bc-401e-92f0-30e1d99130bc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.832235 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e2fd795f-95bc-401e-92f0-30e1d99130bc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lgqfz\" (UID: \"e2fd795f-95bc-401e-92f0-30e1d99130bc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.846885 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2fd795f-95bc-401e-92f0-30e1d99130bc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lgqfz\" (UID: \"e2fd795f-95bc-401e-92f0-30e1d99130bc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.850406 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e2fd795f-95bc-401e-92f0-30e1d99130bc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lgqfz\" (UID: \"e2fd795f-95bc-401e-92f0-30e1d99130bc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" Dec 03 06:28:55 crc kubenswrapper[4818]: I1203 06:28:55.957150 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" Dec 03 06:28:56 crc kubenswrapper[4818]: I1203 06:28:56.230513 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" event={"ID":"e2fd795f-95bc-401e-92f0-30e1d99130bc","Type":"ContainerStarted","Data":"5529babab561dff9e1c61a3c9229463847d291c3e4b99ab22bf9f206f4c67d9e"} Dec 03 06:28:56 crc kubenswrapper[4818]: I1203 06:28:56.230874 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" event={"ID":"e2fd795f-95bc-401e-92f0-30e1d99130bc","Type":"ContainerStarted","Data":"6599c2fe7f6f77c14a721974e3d8c19db1742352cce33b1bacf9be4bf18c39d9"} Dec 03 06:28:56 crc kubenswrapper[4818]: I1203 06:28:56.243925 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=44.243905792 podStartE2EDuration="44.243905792s" podCreationTimestamp="2025-12-03 06:28:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:28:55.702833158 +0000 UTC m=+93.394441910" watchObservedRunningTime="2025-12-03 06:28:56.243905792 +0000 UTC m=+93.935514544" Dec 03 06:28:56 crc kubenswrapper[4818]: I1203 06:28:56.736715 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:56 crc kubenswrapper[4818]: E1203 06:28:56.736881 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:57 crc kubenswrapper[4818]: I1203 06:28:57.737249 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:57 crc kubenswrapper[4818]: I1203 06:28:57.737255 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:57 crc kubenswrapper[4818]: E1203 06:28:57.737409 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:57 crc kubenswrapper[4818]: E1203 06:28:57.737608 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:28:57 crc kubenswrapper[4818]: I1203 06:28:57.737952 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:57 crc kubenswrapper[4818]: E1203 06:28:57.738149 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:58 crc kubenswrapper[4818]: I1203 06:28:58.736764 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:28:58 crc kubenswrapper[4818]: E1203 06:28:58.736921 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:28:59 crc kubenswrapper[4818]: I1203 06:28:59.736926 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:28:59 crc kubenswrapper[4818]: I1203 06:28:59.737075 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:28:59 crc kubenswrapper[4818]: E1203 06:28:59.737156 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:28:59 crc kubenswrapper[4818]: I1203 06:28:59.737188 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:28:59 crc kubenswrapper[4818]: E1203 06:28:59.737416 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:28:59 crc kubenswrapper[4818]: E1203 06:28:59.737491 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:00 crc kubenswrapper[4818]: I1203 06:29:00.737037 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:00 crc kubenswrapper[4818]: E1203 06:29:00.737232 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:01 crc kubenswrapper[4818]: I1203 06:29:01.736885 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:01 crc kubenswrapper[4818]: I1203 06:29:01.737031 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:01 crc kubenswrapper[4818]: E1203 06:29:01.737079 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:01 crc kubenswrapper[4818]: I1203 06:29:01.737181 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:01 crc kubenswrapper[4818]: E1203 06:29:01.738274 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:01 crc kubenswrapper[4818]: E1203 06:29:01.738309 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:01 crc kubenswrapper[4818]: I1203 06:29:01.738442 4818 scope.go:117] "RemoveContainer" containerID="0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024" Dec 03 06:29:01 crc kubenswrapper[4818]: E1203 06:29:01.738969 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" Dec 03 06:29:02 crc kubenswrapper[4818]: I1203 06:29:02.736634 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:02 crc kubenswrapper[4818]: E1203 06:29:02.739022 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:03 crc kubenswrapper[4818]: I1203 06:29:03.737074 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:03 crc kubenswrapper[4818]: I1203 06:29:03.737120 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:03 crc kubenswrapper[4818]: I1203 06:29:03.737095 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:03 crc kubenswrapper[4818]: E1203 06:29:03.737294 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:03 crc kubenswrapper[4818]: E1203 06:29:03.737434 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:03 crc kubenswrapper[4818]: E1203 06:29:03.737706 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:04 crc kubenswrapper[4818]: I1203 06:29:04.333216 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs\") pod \"network-metrics-daemon-gq8r2\" (UID: \"2d8af736-67f3-4db0-a7a5-a24af4222b86\") " pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:04 crc kubenswrapper[4818]: E1203 06:29:04.333376 4818 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:29:04 crc kubenswrapper[4818]: E1203 06:29:04.333440 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs podName:2d8af736-67f3-4db0-a7a5-a24af4222b86 nodeName:}" failed. No retries permitted until 2025-12-03 06:30:08.333425695 +0000 UTC m=+166.025034447 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs") pod "network-metrics-daemon-gq8r2" (UID: "2d8af736-67f3-4db0-a7a5-a24af4222b86") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 06:29:04 crc kubenswrapper[4818]: I1203 06:29:04.737433 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:04 crc kubenswrapper[4818]: E1203 06:29:04.737609 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:05 crc kubenswrapper[4818]: I1203 06:29:05.736780 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:05 crc kubenswrapper[4818]: E1203 06:29:05.736896 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:05 crc kubenswrapper[4818]: I1203 06:29:05.736793 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:05 crc kubenswrapper[4818]: I1203 06:29:05.736975 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:05 crc kubenswrapper[4818]: E1203 06:29:05.737110 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:05 crc kubenswrapper[4818]: E1203 06:29:05.737384 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:06 crc kubenswrapper[4818]: I1203 06:29:06.736595 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:06 crc kubenswrapper[4818]: E1203 06:29:06.736805 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:07 crc kubenswrapper[4818]: I1203 06:29:07.737161 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:07 crc kubenswrapper[4818]: I1203 06:29:07.737241 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:07 crc kubenswrapper[4818]: I1203 06:29:07.737272 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:07 crc kubenswrapper[4818]: E1203 06:29:07.738496 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:07 crc kubenswrapper[4818]: E1203 06:29:07.738598 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:07 crc kubenswrapper[4818]: E1203 06:29:07.738897 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:08 crc kubenswrapper[4818]: I1203 06:29:08.736774 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:08 crc kubenswrapper[4818]: E1203 06:29:08.737168 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:09 crc kubenswrapper[4818]: I1203 06:29:09.736691 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:09 crc kubenswrapper[4818]: I1203 06:29:09.736791 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:09 crc kubenswrapper[4818]: E1203 06:29:09.736853 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:09 crc kubenswrapper[4818]: I1203 06:29:09.736956 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:09 crc kubenswrapper[4818]: E1203 06:29:09.737088 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:09 crc kubenswrapper[4818]: E1203 06:29:09.737513 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:10 crc kubenswrapper[4818]: I1203 06:29:10.736706 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:10 crc kubenswrapper[4818]: E1203 06:29:10.737091 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:11 crc kubenswrapper[4818]: I1203 06:29:11.736773 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:11 crc kubenswrapper[4818]: I1203 06:29:11.736798 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:11 crc kubenswrapper[4818]: I1203 06:29:11.736852 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:11 crc kubenswrapper[4818]: E1203 06:29:11.736930 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:11 crc kubenswrapper[4818]: E1203 06:29:11.736995 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:11 crc kubenswrapper[4818]: E1203 06:29:11.737057 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:12 crc kubenswrapper[4818]: I1203 06:29:12.737607 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:12 crc kubenswrapper[4818]: E1203 06:29:12.738368 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:13 crc kubenswrapper[4818]: I1203 06:29:13.736481 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:13 crc kubenswrapper[4818]: I1203 06:29:13.736576 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:13 crc kubenswrapper[4818]: I1203 06:29:13.736576 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:13 crc kubenswrapper[4818]: E1203 06:29:13.736681 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:13 crc kubenswrapper[4818]: E1203 06:29:13.736919 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:13 crc kubenswrapper[4818]: E1203 06:29:13.736962 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:14 crc kubenswrapper[4818]: I1203 06:29:14.737420 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:14 crc kubenswrapper[4818]: E1203 06:29:14.737628 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:14 crc kubenswrapper[4818]: I1203 06:29:14.737801 4818 scope.go:117] "RemoveContainer" containerID="0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024" Dec 03 06:29:14 crc kubenswrapper[4818]: E1203 06:29:14.737974 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" Dec 03 06:29:15 crc kubenswrapper[4818]: I1203 06:29:15.737295 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:15 crc kubenswrapper[4818]: E1203 06:29:15.737771 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:15 crc kubenswrapper[4818]: I1203 06:29:15.737392 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:15 crc kubenswrapper[4818]: E1203 06:29:15.738649 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:15 crc kubenswrapper[4818]: I1203 06:29:15.737360 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:15 crc kubenswrapper[4818]: E1203 06:29:15.738898 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:16 crc kubenswrapper[4818]: I1203 06:29:16.736850 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:16 crc kubenswrapper[4818]: E1203 06:29:16.737032 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:17 crc kubenswrapper[4818]: I1203 06:29:17.737476 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:17 crc kubenswrapper[4818]: I1203 06:29:17.737477 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:17 crc kubenswrapper[4818]: I1203 06:29:17.737471 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:17 crc kubenswrapper[4818]: E1203 06:29:17.738354 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:17 crc kubenswrapper[4818]: E1203 06:29:17.738412 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:17 crc kubenswrapper[4818]: E1203 06:29:17.738183 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:18 crc kubenswrapper[4818]: I1203 06:29:18.736535 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:18 crc kubenswrapper[4818]: E1203 06:29:18.737046 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:19 crc kubenswrapper[4818]: I1203 06:29:19.306771 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ctpzs_7b25b836-b639-4111-bcea-af7cc791ea32/kube-multus/1.log" Dec 03 06:29:19 crc kubenswrapper[4818]: I1203 06:29:19.307465 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ctpzs_7b25b836-b639-4111-bcea-af7cc791ea32/kube-multus/0.log" Dec 03 06:29:19 crc kubenswrapper[4818]: I1203 06:29:19.307532 4818 generic.go:334] "Generic (PLEG): container finished" podID="7b25b836-b639-4111-bcea-af7cc791ea32" containerID="8b92f0662c82eef1fb0023107da93fead648febd6e3880b77ad264ced57de7be" exitCode=1 Dec 03 06:29:19 crc kubenswrapper[4818]: I1203 06:29:19.307578 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ctpzs" event={"ID":"7b25b836-b639-4111-bcea-af7cc791ea32","Type":"ContainerDied","Data":"8b92f0662c82eef1fb0023107da93fead648febd6e3880b77ad264ced57de7be"} Dec 03 06:29:19 crc kubenswrapper[4818]: I1203 06:29:19.307629 4818 scope.go:117] "RemoveContainer" containerID="3b045c99249eef21478fe822f75f6a3655b17d31bc8363924c99a7a4527a9b56" Dec 03 06:29:19 crc kubenswrapper[4818]: I1203 06:29:19.308216 4818 scope.go:117] "RemoveContainer" containerID="8b92f0662c82eef1fb0023107da93fead648febd6e3880b77ad264ced57de7be" Dec 03 06:29:19 crc kubenswrapper[4818]: E1203 06:29:19.308487 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-ctpzs_openshift-multus(7b25b836-b639-4111-bcea-af7cc791ea32)\"" pod="openshift-multus/multus-ctpzs" podUID="7b25b836-b639-4111-bcea-af7cc791ea32" Dec 03 06:29:19 crc kubenswrapper[4818]: I1203 06:29:19.341622 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lgqfz" podStartSLOduration=93.34160342 podStartE2EDuration="1m33.34160342s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:28:56.244278511 +0000 UTC m=+93.935887273" watchObservedRunningTime="2025-12-03 06:29:19.34160342 +0000 UTC m=+117.033212192" Dec 03 06:29:19 crc kubenswrapper[4818]: I1203 06:29:19.737072 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:19 crc kubenswrapper[4818]: I1203 06:29:19.737157 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:19 crc kubenswrapper[4818]: I1203 06:29:19.737188 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:19 crc kubenswrapper[4818]: E1203 06:29:19.737307 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:19 crc kubenswrapper[4818]: E1203 06:29:19.737532 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:19 crc kubenswrapper[4818]: E1203 06:29:19.737618 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:20 crc kubenswrapper[4818]: I1203 06:29:20.314109 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ctpzs_7b25b836-b639-4111-bcea-af7cc791ea32/kube-multus/1.log" Dec 03 06:29:20 crc kubenswrapper[4818]: I1203 06:29:20.737206 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:20 crc kubenswrapper[4818]: E1203 06:29:20.737343 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:21 crc kubenswrapper[4818]: I1203 06:29:21.736651 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:21 crc kubenswrapper[4818]: I1203 06:29:21.736703 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:21 crc kubenswrapper[4818]: I1203 06:29:21.736728 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:21 crc kubenswrapper[4818]: E1203 06:29:21.736882 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:21 crc kubenswrapper[4818]: E1203 06:29:21.736997 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:21 crc kubenswrapper[4818]: E1203 06:29:21.737063 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:22 crc kubenswrapper[4818]: E1203 06:29:22.724396 4818 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 03 06:29:22 crc kubenswrapper[4818]: I1203 06:29:22.737445 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:22 crc kubenswrapper[4818]: E1203 06:29:22.739160 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:22 crc kubenswrapper[4818]: E1203 06:29:22.836094 4818 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 06:29:23 crc kubenswrapper[4818]: I1203 06:29:23.737165 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:23 crc kubenswrapper[4818]: I1203 06:29:23.737164 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:23 crc kubenswrapper[4818]: E1203 06:29:23.737361 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:23 crc kubenswrapper[4818]: E1203 06:29:23.737392 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:23 crc kubenswrapper[4818]: I1203 06:29:23.737178 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:23 crc kubenswrapper[4818]: E1203 06:29:23.737454 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:24 crc kubenswrapper[4818]: I1203 06:29:24.737009 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:24 crc kubenswrapper[4818]: E1203 06:29:24.737197 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:25 crc kubenswrapper[4818]: I1203 06:29:25.736895 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:25 crc kubenswrapper[4818]: I1203 06:29:25.736942 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:25 crc kubenswrapper[4818]: E1203 06:29:25.738280 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:25 crc kubenswrapper[4818]: I1203 06:29:25.736958 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:25 crc kubenswrapper[4818]: E1203 06:29:25.738410 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:25 crc kubenswrapper[4818]: E1203 06:29:25.738448 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:26 crc kubenswrapper[4818]: I1203 06:29:26.737462 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:26 crc kubenswrapper[4818]: E1203 06:29:26.737802 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:27 crc kubenswrapper[4818]: I1203 06:29:27.736987 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:27 crc kubenswrapper[4818]: I1203 06:29:27.737022 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:27 crc kubenswrapper[4818]: I1203 06:29:27.737151 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:27 crc kubenswrapper[4818]: E1203 06:29:27.737681 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:27 crc kubenswrapper[4818]: I1203 06:29:27.737922 4818 scope.go:117] "RemoveContainer" containerID="0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024" Dec 03 06:29:27 crc kubenswrapper[4818]: E1203 06:29:27.738047 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:27 crc kubenswrapper[4818]: E1203 06:29:27.738084 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ktm8k_openshift-ovn-kubernetes(a1e30fe6-0b94-4c95-b941-c0623729e123)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" Dec 03 06:29:27 crc kubenswrapper[4818]: E1203 06:29:27.738312 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:27 crc kubenswrapper[4818]: E1203 06:29:27.837063 4818 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 06:29:28 crc kubenswrapper[4818]: I1203 06:29:28.737630 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:28 crc kubenswrapper[4818]: E1203 06:29:28.738286 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:29 crc kubenswrapper[4818]: I1203 06:29:29.736628 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:29 crc kubenswrapper[4818]: I1203 06:29:29.736679 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:29 crc kubenswrapper[4818]: I1203 06:29:29.736737 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:29 crc kubenswrapper[4818]: E1203 06:29:29.736772 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:29 crc kubenswrapper[4818]: E1203 06:29:29.736936 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:29 crc kubenswrapper[4818]: E1203 06:29:29.737000 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:30 crc kubenswrapper[4818]: I1203 06:29:30.736649 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:30 crc kubenswrapper[4818]: E1203 06:29:30.736896 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:30 crc kubenswrapper[4818]: I1203 06:29:30.736972 4818 scope.go:117] "RemoveContainer" containerID="8b92f0662c82eef1fb0023107da93fead648febd6e3880b77ad264ced57de7be" Dec 03 06:29:31 crc kubenswrapper[4818]: I1203 06:29:31.353734 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ctpzs_7b25b836-b639-4111-bcea-af7cc791ea32/kube-multus/1.log" Dec 03 06:29:31 crc kubenswrapper[4818]: I1203 06:29:31.353802 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ctpzs" event={"ID":"7b25b836-b639-4111-bcea-af7cc791ea32","Type":"ContainerStarted","Data":"3860840e364d23f4d54acf2e87b4f131ee6d2bbd80175613d39d96953478088c"} Dec 03 06:29:31 crc kubenswrapper[4818]: I1203 06:29:31.737090 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:31 crc kubenswrapper[4818]: I1203 06:29:31.737137 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:31 crc kubenswrapper[4818]: I1203 06:29:31.737233 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:31 crc kubenswrapper[4818]: E1203 06:29:31.737484 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:31 crc kubenswrapper[4818]: E1203 06:29:31.737625 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:31 crc kubenswrapper[4818]: E1203 06:29:31.737702 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:32 crc kubenswrapper[4818]: I1203 06:29:32.737146 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:32 crc kubenswrapper[4818]: E1203 06:29:32.739736 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:32 crc kubenswrapper[4818]: E1203 06:29:32.837701 4818 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 06:29:33 crc kubenswrapper[4818]: I1203 06:29:33.737042 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:33 crc kubenswrapper[4818]: I1203 06:29:33.737140 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:33 crc kubenswrapper[4818]: E1203 06:29:33.737193 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:33 crc kubenswrapper[4818]: I1203 06:29:33.737264 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:33 crc kubenswrapper[4818]: E1203 06:29:33.737639 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:33 crc kubenswrapper[4818]: E1203 06:29:33.737765 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:34 crc kubenswrapper[4818]: I1203 06:29:34.737583 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:34 crc kubenswrapper[4818]: E1203 06:29:34.737753 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:35 crc kubenswrapper[4818]: I1203 06:29:35.736873 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:35 crc kubenswrapper[4818]: I1203 06:29:35.736928 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:35 crc kubenswrapper[4818]: I1203 06:29:35.736975 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:35 crc kubenswrapper[4818]: E1203 06:29:35.737015 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:35 crc kubenswrapper[4818]: E1203 06:29:35.737180 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:35 crc kubenswrapper[4818]: E1203 06:29:35.737224 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:36 crc kubenswrapper[4818]: I1203 06:29:36.736928 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:36 crc kubenswrapper[4818]: E1203 06:29:36.737088 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:37 crc kubenswrapper[4818]: I1203 06:29:37.737306 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:37 crc kubenswrapper[4818]: I1203 06:29:37.737375 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:37 crc kubenswrapper[4818]: E1203 06:29:37.737462 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:37 crc kubenswrapper[4818]: I1203 06:29:37.737477 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:37 crc kubenswrapper[4818]: E1203 06:29:37.737607 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:37 crc kubenswrapper[4818]: E1203 06:29:37.737669 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:37 crc kubenswrapper[4818]: E1203 06:29:37.838808 4818 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 06:29:38 crc kubenswrapper[4818]: I1203 06:29:38.736838 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:38 crc kubenswrapper[4818]: E1203 06:29:38.737007 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:39 crc kubenswrapper[4818]: I1203 06:29:39.737473 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:39 crc kubenswrapper[4818]: I1203 06:29:39.737560 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:39 crc kubenswrapper[4818]: E1203 06:29:39.737628 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:39 crc kubenswrapper[4818]: E1203 06:29:39.737702 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:39 crc kubenswrapper[4818]: I1203 06:29:39.738200 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:39 crc kubenswrapper[4818]: E1203 06:29:39.738367 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:40 crc kubenswrapper[4818]: I1203 06:29:40.737155 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:40 crc kubenswrapper[4818]: E1203 06:29:40.737319 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:41 crc kubenswrapper[4818]: I1203 06:29:41.736875 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:41 crc kubenswrapper[4818]: I1203 06:29:41.736928 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:41 crc kubenswrapper[4818]: I1203 06:29:41.736893 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:41 crc kubenswrapper[4818]: E1203 06:29:41.737087 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:41 crc kubenswrapper[4818]: E1203 06:29:41.737233 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:41 crc kubenswrapper[4818]: E1203 06:29:41.737546 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:42 crc kubenswrapper[4818]: I1203 06:29:42.737396 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:42 crc kubenswrapper[4818]: E1203 06:29:42.738793 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:42 crc kubenswrapper[4818]: I1203 06:29:42.739577 4818 scope.go:117] "RemoveContainer" containerID="0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024" Dec 03 06:29:42 crc kubenswrapper[4818]: E1203 06:29:42.840043 4818 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 06:29:43 crc kubenswrapper[4818]: I1203 06:29:43.425968 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovnkube-controller/3.log" Dec 03 06:29:43 crc kubenswrapper[4818]: I1203 06:29:43.428276 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerStarted","Data":"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc"} Dec 03 06:29:43 crc kubenswrapper[4818]: I1203 06:29:43.429148 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:29:43 crc kubenswrapper[4818]: I1203 06:29:43.695133 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podStartSLOduration=117.695107964 podStartE2EDuration="1m57.695107964s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:29:43.462384243 +0000 UTC m=+141.153992995" watchObservedRunningTime="2025-12-03 06:29:43.695107964 +0000 UTC m=+141.386716726" Dec 03 06:29:43 crc kubenswrapper[4818]: I1203 06:29:43.696319 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gq8r2"] Dec 03 06:29:43 crc kubenswrapper[4818]: I1203 06:29:43.696447 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:43 crc kubenswrapper[4818]: E1203 06:29:43.696578 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:43 crc kubenswrapper[4818]: I1203 06:29:43.737044 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:43 crc kubenswrapper[4818]: I1203 06:29:43.737116 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:43 crc kubenswrapper[4818]: E1203 06:29:43.737187 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:43 crc kubenswrapper[4818]: E1203 06:29:43.737311 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:44 crc kubenswrapper[4818]: I1203 06:29:44.737132 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:44 crc kubenswrapper[4818]: E1203 06:29:44.737330 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:45 crc kubenswrapper[4818]: I1203 06:29:45.737036 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:45 crc kubenswrapper[4818]: I1203 06:29:45.737097 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:45 crc kubenswrapper[4818]: I1203 06:29:45.737227 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:45 crc kubenswrapper[4818]: E1203 06:29:45.737669 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:45 crc kubenswrapper[4818]: E1203 06:29:45.737796 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:45 crc kubenswrapper[4818]: E1203 06:29:45.737926 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:46 crc kubenswrapper[4818]: I1203 06:29:46.737378 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:46 crc kubenswrapper[4818]: E1203 06:29:46.737588 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:29:47 crc kubenswrapper[4818]: I1203 06:29:47.736882 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:47 crc kubenswrapper[4818]: I1203 06:29:47.736994 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:47 crc kubenswrapper[4818]: E1203 06:29:47.737089 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:29:47 crc kubenswrapper[4818]: E1203 06:29:47.737162 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:29:47 crc kubenswrapper[4818]: I1203 06:29:47.736917 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:47 crc kubenswrapper[4818]: E1203 06:29:47.737272 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-gq8r2" podUID="2d8af736-67f3-4db0-a7a5-a24af4222b86" Dec 03 06:29:48 crc kubenswrapper[4818]: I1203 06:29:48.520589 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:29:48 crc kubenswrapper[4818]: E1203 06:29:48.520935 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:31:50.520895035 +0000 UTC m=+268.212503787 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:29:48 crc kubenswrapper[4818]: I1203 06:29:48.622029 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:48 crc kubenswrapper[4818]: I1203 06:29:48.622107 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:48 crc kubenswrapper[4818]: I1203 06:29:48.622151 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:48 crc kubenswrapper[4818]: I1203 06:29:48.622241 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:48 crc kubenswrapper[4818]: E1203 06:29:48.622314 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:29:48 crc kubenswrapper[4818]: E1203 06:29:48.622339 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:29:48 crc kubenswrapper[4818]: E1203 06:29:48.622333 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:29:48 crc kubenswrapper[4818]: E1203 06:29:48.622473 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:31:50.622440188 +0000 UTC m=+268.314048940 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 06:29:48 crc kubenswrapper[4818]: E1203 06:29:48.622354 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:29:48 crc kubenswrapper[4818]: E1203 06:29:48.622349 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:29:48 crc kubenswrapper[4818]: E1203 06:29:48.622722 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:31:50.622667193 +0000 UTC m=+268.314276145 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:29:48 crc kubenswrapper[4818]: E1203 06:29:48.622755 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:31:50.622745845 +0000 UTC m=+268.314354597 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 06:29:48 crc kubenswrapper[4818]: E1203 06:29:48.622487 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 06:29:48 crc kubenswrapper[4818]: E1203 06:29:48.622796 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 06:29:48 crc kubenswrapper[4818]: E1203 06:29:48.622837 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:29:48 crc kubenswrapper[4818]: E1203 06:29:48.622891 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:31:50.622881679 +0000 UTC m=+268.314490661 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 06:29:48 crc kubenswrapper[4818]: I1203 06:29:48.737666 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:29:48 crc kubenswrapper[4818]: I1203 06:29:48.739503 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 06:29:48 crc kubenswrapper[4818]: I1203 06:29:48.741175 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 06:29:49 crc kubenswrapper[4818]: I1203 06:29:49.736893 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:29:49 crc kubenswrapper[4818]: I1203 06:29:49.736974 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:29:49 crc kubenswrapper[4818]: I1203 06:29:49.736973 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:29:49 crc kubenswrapper[4818]: I1203 06:29:49.740333 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 06:29:49 crc kubenswrapper[4818]: I1203 06:29:49.740413 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 06:29:49 crc kubenswrapper[4818]: I1203 06:29:49.740936 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 06:29:49 crc kubenswrapper[4818]: I1203 06:29:49.744426 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.764034 4818 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.796920 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jf227"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.798598 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.805079 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cwhmk"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.805407 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.805592 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.805962 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.806612 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.808188 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.810172 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j8d4t"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.810685 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j8d4t" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.810735 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rl64s"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.811016 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.812185 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.812542 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.813022 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.813194 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.813292 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.813326 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.813408 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.813451 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.813418 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.816607 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qnxnn"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.817135 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tx2n9"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.817396 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.817805 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.818647 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.818870 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tx2n9" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.821607 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2sn7n"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.821909 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.821933 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.821998 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.822068 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.822149 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.822261 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.822455 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-2sn7n" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.822829 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.822890 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.822966 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.822975 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.823135 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.823316 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.823330 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gt4k9"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.823412 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.823680 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gt4k9" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.823807 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.824315 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.824317 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.826881 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6mbg"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.827471 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6mbg" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.827567 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-pvb2q"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.828082 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-pvb2q" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.829497 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.830398 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fv5lc"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.830915 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zg4cb"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.831278 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.831334 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.832427 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.832707 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.832428 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.833314 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.833541 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.833732 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.833339 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.834348 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.834520 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.836200 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.836210 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.836466 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.836551 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.836701 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.836767 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.836961 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.837089 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.849625 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.850011 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.850242 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.849662 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.836578 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.866564 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-sccrw"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.876513 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.877251 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j25z9"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.877608 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.877621 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.877771 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.877901 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.878173 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.878777 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.879327 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.880026 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.880319 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j25z9" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.880981 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.886585 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.887249 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.887337 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.887843 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.889786 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.890094 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.892972 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.893955 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.894327 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.894500 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.894640 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.894797 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.894923 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.895000 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.895087 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.895339 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.895419 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.895493 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.895568 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.895649 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.895736 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.895849 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.895934 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.896014 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.896088 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.896169 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.896262 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.896361 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.896448 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.896542 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.896633 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.896733 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.896750 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.896855 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.897027 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.897344 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.897418 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.897450 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.897555 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.897634 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.897715 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.897788 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.897889 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.898028 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.897894 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.898416 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.898930 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.899105 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.899286 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6h7lb"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.899836 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.901509 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.904290 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.908710 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-fz577"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909209 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s8mcw"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909439 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a406cec-8381-4a82-8d46-3ababcda72b4-config\") pod \"route-controller-manager-6576b87f9c-5jw58\" (UID: \"6a406cec-8381-4a82-8d46-3ababcda72b4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909493 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0122f472-9da5-40c1-aa0a-ab6219543b1d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909513 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s8mcw" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909522 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrbwf\" (UniqueName: \"kubernetes.io/projected/8ec6e899-8380-4812-8f65-6fd72db12939-kube-api-access-lrbwf\") pod \"machine-api-operator-5694c8668f-jf227\" (UID: \"8ec6e899-8380-4812-8f65-6fd72db12939\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909562 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0122f472-9da5-40c1-aa0a-ab6219543b1d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909585 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2m55\" (UniqueName: \"kubernetes.io/projected/0122f472-9da5-40c1-aa0a-ab6219543b1d-kube-api-access-w2m55\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909612 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d2wq\" (UniqueName: \"kubernetes.io/projected/6a406cec-8381-4a82-8d46-3ababcda72b4-kube-api-access-8d2wq\") pod \"route-controller-manager-6576b87f9c-5jw58\" (UID: \"6a406cec-8381-4a82-8d46-3ababcda72b4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909633 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0122f472-9da5-40c1-aa0a-ab6219543b1d-audit-policies\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909654 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0122f472-9da5-40c1-aa0a-ab6219543b1d-serving-cert\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909675 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0122f472-9da5-40c1-aa0a-ab6219543b1d-encryption-config\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909699 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-client-ca\") pod \"controller-manager-879f6c89f-cwhmk\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909724 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-cwhmk\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909744 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8ec6e899-8380-4812-8f65-6fd72db12939-images\") pod \"machine-api-operator-5694c8668f-jf227\" (UID: \"8ec6e899-8380-4812-8f65-6fd72db12939\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909760 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909772 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a406cec-8381-4a82-8d46-3ababcda72b4-serving-cert\") pod \"route-controller-manager-6576b87f9c-5jw58\" (UID: \"6a406cec-8381-4a82-8d46-3ababcda72b4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909795 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-config\") pod \"controller-manager-879f6c89f-cwhmk\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909846 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0122f472-9da5-40c1-aa0a-ab6219543b1d-audit-dir\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909888 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr2fg\" (UniqueName: \"kubernetes.io/projected/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-kube-api-access-wr2fg\") pod \"controller-manager-879f6c89f-cwhmk\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.909913 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0122f472-9da5-40c1-aa0a-ab6219543b1d-etcd-client\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.910069 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ec6e899-8380-4812-8f65-6fd72db12939-config\") pod \"machine-api-operator-5694c8668f-jf227\" (UID: \"8ec6e899-8380-4812-8f65-6fd72db12939\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.910201 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a406cec-8381-4a82-8d46-3ababcda72b4-client-ca\") pod \"route-controller-manager-6576b87f9c-5jw58\" (UID: \"6a406cec-8381-4a82-8d46-3ababcda72b4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.910308 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8ec6e899-8380-4812-8f65-6fd72db12939-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jf227\" (UID: \"8ec6e899-8380-4812-8f65-6fd72db12939\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.910877 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.910873 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-serving-cert\") pod \"controller-manager-879f6c89f-cwhmk\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.932713 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.941760 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dh7z8"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.945386 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8x7mv"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.945570 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.945644 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dh7z8" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.947424 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.947955 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8x7mv" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.947564 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.947766 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.949263 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jf227"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.950833 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.951254 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.951397 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jvbsr"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.954414 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jvbsr" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.955335 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cwhmk"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.957217 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rl64s"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.960601 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.964415 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.965164 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.966782 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ltb69"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.967513 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ltb69" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.967743 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4dh8d"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.968191 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.971195 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.971587 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.972459 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-596cp"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.973041 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-596cp" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.973671 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.973888 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.974641 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.975201 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.976299 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j8d4t"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.977972 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.979870 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-s8vm2"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.980729 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6glnm"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.981236 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-s8vm2" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.981349 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6glnm" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.981915 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.982766 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-phwq8"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.983341 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-phwq8" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.984349 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tx2n9"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.985594 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.987580 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-l6nvl"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.988306 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-l6nvl" Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.989249 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.994365 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2sn7n"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.996068 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s8mcw"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.998605 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-pvb2q"] Dec 03 06:29:56 crc kubenswrapper[4818]: I1203 06:29:56.999893 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dh7z8"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.000416 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.001278 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j25z9"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.002872 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.004204 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.005470 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6mbg"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.007388 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011487 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-client-ca\") pod \"controller-manager-879f6c89f-cwhmk\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011527 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011550 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-cwhmk\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011571 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a406cec-8381-4a82-8d46-3ababcda72b4-serving-cert\") pod \"route-controller-manager-6576b87f9c-5jw58\" (UID: \"6a406cec-8381-4a82-8d46-3ababcda72b4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011590 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8ec6e899-8380-4812-8f65-6fd72db12939-images\") pod \"machine-api-operator-5694c8668f-jf227\" (UID: \"8ec6e899-8380-4812-8f65-6fd72db12939\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011611 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d7f12c1b-f504-4155-bd29-bcb6ab57b646-audit-dir\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011634 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-config\") pod \"controller-manager-879f6c89f-cwhmk\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011656 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011677 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011704 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0122f472-9da5-40c1-aa0a-ab6219543b1d-audit-dir\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011722 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011741 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr2fg\" (UniqueName: \"kubernetes.io/projected/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-kube-api-access-wr2fg\") pod \"controller-manager-879f6c89f-cwhmk\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011767 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ec6e899-8380-4812-8f65-6fd72db12939-config\") pod \"machine-api-operator-5694c8668f-jf227\" (UID: \"8ec6e899-8380-4812-8f65-6fd72db12939\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011789 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0122f472-9da5-40c1-aa0a-ab6219543b1d-etcd-client\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011809 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011870 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8ec6e899-8380-4812-8f65-6fd72db12939-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jf227\" (UID: \"8ec6e899-8380-4812-8f65-6fd72db12939\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011887 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-serving-cert\") pod \"controller-manager-879f6c89f-cwhmk\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011904 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011924 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a406cec-8381-4a82-8d46-3ababcda72b4-client-ca\") pod \"route-controller-manager-6576b87f9c-5jw58\" (UID: \"6a406cec-8381-4a82-8d46-3ababcda72b4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011955 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011981 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a406cec-8381-4a82-8d46-3ababcda72b4-config\") pod \"route-controller-manager-6576b87f9c-5jw58\" (UID: \"6a406cec-8381-4a82-8d46-3ababcda72b4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.011998 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012014 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012029 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvlzn\" (UniqueName: \"kubernetes.io/projected/d7f12c1b-f504-4155-bd29-bcb6ab57b646-kube-api-access-pvlzn\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012047 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012066 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0122f472-9da5-40c1-aa0a-ab6219543b1d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012122 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85223238-8e03-42bb-8790-8887e80d0e52-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tx2n9\" (UID: \"85223238-8e03-42bb-8790-8887e80d0e52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tx2n9" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012179 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrbwf\" (UniqueName: \"kubernetes.io/projected/8ec6e899-8380-4812-8f65-6fd72db12939-kube-api-access-lrbwf\") pod \"machine-api-operator-5694c8668f-jf227\" (UID: \"8ec6e899-8380-4812-8f65-6fd72db12939\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012212 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012237 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-audit-policies\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012257 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85223238-8e03-42bb-8790-8887e80d0e52-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tx2n9\" (UID: \"85223238-8e03-42bb-8790-8887e80d0e52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tx2n9" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012307 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0122f472-9da5-40c1-aa0a-ab6219543b1d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012328 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d2wq\" (UniqueName: \"kubernetes.io/projected/6a406cec-8381-4a82-8d46-3ababcda72b4-kube-api-access-8d2wq\") pod \"route-controller-manager-6576b87f9c-5jw58\" (UID: \"6a406cec-8381-4a82-8d46-3ababcda72b4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012347 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0122f472-9da5-40c1-aa0a-ab6219543b1d-audit-policies\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012354 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-client-ca\") pod \"controller-manager-879f6c89f-cwhmk\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012364 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2m55\" (UniqueName: \"kubernetes.io/projected/0122f472-9da5-40c1-aa0a-ab6219543b1d-kube-api-access-w2m55\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012398 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0122f472-9da5-40c1-aa0a-ab6219543b1d-serving-cert\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012416 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0122f472-9da5-40c1-aa0a-ab6219543b1d-encryption-config\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012435 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65854\" (UniqueName: \"kubernetes.io/projected/85223238-8e03-42bb-8790-8887e80d0e52-kube-api-access-65854\") pod \"openshift-apiserver-operator-796bbdcf4f-tx2n9\" (UID: \"85223238-8e03-42bb-8790-8887e80d0e52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tx2n9" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.012501 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0122f472-9da5-40c1-aa0a-ab6219543b1d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.013304 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-cwhmk\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.013348 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ec6e899-8380-4812-8f65-6fd72db12939-config\") pod \"machine-api-operator-5694c8668f-jf227\" (UID: \"8ec6e899-8380-4812-8f65-6fd72db12939\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.013433 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0122f472-9da5-40c1-aa0a-ab6219543b1d-audit-dir\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.013900 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-config\") pod \"controller-manager-879f6c89f-cwhmk\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.014037 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gt4k9"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.014199 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8ec6e899-8380-4812-8f65-6fd72db12939-images\") pod \"machine-api-operator-5694c8668f-jf227\" (UID: \"8ec6e899-8380-4812-8f65-6fd72db12939\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.014236 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0122f472-9da5-40c1-aa0a-ab6219543b1d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.014384 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0122f472-9da5-40c1-aa0a-ab6219543b1d-audit-policies\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.014614 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qnxnn"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.015337 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a406cec-8381-4a82-8d46-3ababcda72b4-client-ca\") pod \"route-controller-manager-6576b87f9c-5jw58\" (UID: \"6a406cec-8381-4a82-8d46-3ababcda72b4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.015629 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a406cec-8381-4a82-8d46-3ababcda72b4-config\") pod \"route-controller-manager-6576b87f9c-5jw58\" (UID: \"6a406cec-8381-4a82-8d46-3ababcda72b4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.018201 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4dh8d"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.021242 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-serving-cert\") pod \"controller-manager-879f6c89f-cwhmk\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.021978 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0122f472-9da5-40c1-aa0a-ab6219543b1d-etcd-client\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.023423 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a406cec-8381-4a82-8d46-3ababcda72b4-serving-cert\") pod \"route-controller-manager-6576b87f9c-5jw58\" (UID: \"6a406cec-8381-4a82-8d46-3ababcda72b4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.023908 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zg4cb"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.025159 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.025270 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0122f472-9da5-40c1-aa0a-ab6219543b1d-serving-cert\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.025762 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.027003 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-sccrw"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.028316 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fv5lc"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.033586 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-596cp"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.035559 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.037494 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jvbsr"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.041497 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0122f472-9da5-40c1-aa0a-ab6219543b1d-encryption-config\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.041511 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8x7mv"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.041973 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.048359 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-s8vm2"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.048870 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ltb69"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.049530 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/8ec6e899-8380-4812-8f65-6fd72db12939-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jf227\" (UID: \"8ec6e899-8380-4812-8f65-6fd72db12939\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.051616 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6h7lb"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.052348 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6glnm"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.053468 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.054421 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.055507 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.056461 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-nmqf4"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.058094 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-phwq8"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.058178 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-nmqf4" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.059585 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-nmqf4"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.059732 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-79bwt"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.060401 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-79bwt" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.061262 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.062114 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gkgqw"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.063039 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.064023 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-79bwt"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.065502 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gkgqw"] Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.080741 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.105256 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113063 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113098 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113120 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113149 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113165 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113179 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvlzn\" (UniqueName: \"kubernetes.io/projected/d7f12c1b-f504-4155-bd29-bcb6ab57b646-kube-api-access-pvlzn\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113197 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113214 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85223238-8e03-42bb-8790-8887e80d0e52-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tx2n9\" (UID: \"85223238-8e03-42bb-8790-8887e80d0e52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tx2n9" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113244 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113260 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-audit-policies\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113276 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85223238-8e03-42bb-8790-8887e80d0e52-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tx2n9\" (UID: \"85223238-8e03-42bb-8790-8887e80d0e52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tx2n9" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113311 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65854\" (UniqueName: \"kubernetes.io/projected/85223238-8e03-42bb-8790-8887e80d0e52-kube-api-access-65854\") pod \"openshift-apiserver-operator-796bbdcf4f-tx2n9\" (UID: \"85223238-8e03-42bb-8790-8887e80d0e52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tx2n9" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113329 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113345 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d7f12c1b-f504-4155-bd29-bcb6ab57b646-audit-dir\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113362 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113379 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113404 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.113635 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d7f12c1b-f504-4155-bd29-bcb6ab57b646-audit-dir\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.114387 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.114838 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85223238-8e03-42bb-8790-8887e80d0e52-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tx2n9\" (UID: \"85223238-8e03-42bb-8790-8887e80d0e52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tx2n9" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.115016 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-audit-policies\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.115412 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.115992 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.116151 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.116285 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.116547 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.116723 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85223238-8e03-42bb-8790-8887e80d0e52-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tx2n9\" (UID: \"85223238-8e03-42bb-8790-8887e80d0e52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tx2n9" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.117839 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.117993 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.118876 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.119582 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.119787 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.120714 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.141120 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.162343 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.181389 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.201608 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.241349 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.261678 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.281784 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.301051 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.321780 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.342461 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.361295 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.381445 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.400647 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.421373 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.441785 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.461326 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.481124 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.502208 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.521133 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.543053 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.562050 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.582348 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.601780 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.641633 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.660778 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.681868 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.702414 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.721098 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.742199 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.761310 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.780891 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.801693 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.822073 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.841475 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.863490 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.881320 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.901311 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.921185 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.941505 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.959725 4818 request.go:700] Waited for 1.005019337s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns-operator/secrets?fieldSelector=metadata.name%3Dmetrics-tls&limit=500&resourceVersion=0 Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.961408 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 06:29:57 crc kubenswrapper[4818]: I1203 06:29:57.980742 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.002889 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.021757 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.042224 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.068976 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.080623 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.100834 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.121627 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.141411 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.162171 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.181662 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.208432 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.221555 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.241679 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.261419 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.280887 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.300949 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.321482 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.341421 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.362028 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.380978 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.401599 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.422373 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.441436 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.461964 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.482111 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.501549 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.521598 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.541062 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.561998 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.582100 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.601966 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.621736 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.642466 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.681396 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2m55\" (UniqueName: \"kubernetes.io/projected/0122f472-9da5-40c1-aa0a-ab6219543b1d-kube-api-access-w2m55\") pod \"apiserver-7bbb656c7d-npq7z\" (UID: \"0122f472-9da5-40c1-aa0a-ab6219543b1d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.700499 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr2fg\" (UniqueName: \"kubernetes.io/projected/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-kube-api-access-wr2fg\") pod \"controller-manager-879f6c89f-cwhmk\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.711396 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.722336 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrbwf\" (UniqueName: \"kubernetes.io/projected/8ec6e899-8380-4812-8f65-6fd72db12939-kube-api-access-lrbwf\") pod \"machine-api-operator-5694c8668f-jf227\" (UID: \"8ec6e899-8380-4812-8f65-6fd72db12939\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.738246 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d2wq\" (UniqueName: \"kubernetes.io/projected/6a406cec-8381-4a82-8d46-3ababcda72b4-kube-api-access-8d2wq\") pod \"route-controller-manager-6576b87f9c-5jw58\" (UID: \"6a406cec-8381-4a82-8d46-3ababcda72b4\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.741641 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.761578 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.782019 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.801921 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.821562 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.841423 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.862012 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.881729 4818 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.901122 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.906941 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.923048 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.926313 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cwhmk"] Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.934373 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.956626 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvlzn\" (UniqueName: \"kubernetes.io/projected/d7f12c1b-f504-4155-bd29-bcb6ab57b646-kube-api-access-pvlzn\") pod \"oauth-openshift-558db77b4-fv5lc\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.976438 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65854\" (UniqueName: \"kubernetes.io/projected/85223238-8e03-42bb-8790-8887e80d0e52-kube-api-access-65854\") pod \"openshift-apiserver-operator-796bbdcf4f-tx2n9\" (UID: \"85223238-8e03-42bb-8790-8887e80d0e52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tx2n9" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.977253 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:29:58 crc kubenswrapper[4818]: I1203 06:29:58.980601 4818 request.go:700] Waited for 1.757796248s due to client-side throttling, not priority and fairness, request: PATCH:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver-operator/pods/openshift-apiserver-operator-796bbdcf4f-tx2n9/status Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.037372 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/535884b5-264b-4471-962c-51579eeacab0-registry-certificates\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.037792 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a45cf1f5-9679-484d-920c-01e9a0d24587-config\") pod \"console-operator-58897d9998-2sn7n\" (UID: \"a45cf1f5-9679-484d-920c-01e9a0d24587\") " pod="openshift-console-operator/console-operator-58897d9998-2sn7n" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.037852 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-bound-sa-token\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.037878 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/06af2589-0701-4b16-a31c-eb39fe1c8b99-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-j25z9\" (UID: \"06af2589-0701-4b16-a31c-eb39fe1c8b99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j25z9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.037904 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/05f7659c-85e9-4337-8741-9958f12895d4-metrics-certs\") pod \"router-default-5444994796-fz577\" (UID: \"05f7659c-85e9-4337-8741-9958f12895d4\") " pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.037926 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnmvg\" (UniqueName: \"kubernetes.io/projected/3dbb4222-9650-4663-95bf-bd2fa4f4e678-kube-api-access-cnmvg\") pod \"openshift-config-operator-7777fb866f-h6fkg\" (UID: \"3dbb4222-9650-4663-95bf-bd2fa4f4e678\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.037948 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnlc6\" (UniqueName: \"kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-kube-api-access-gnlc6\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.037970 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k4pkt\" (UID: \"aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.037993 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/39dce7bf-1502-40fa-9d9c-41378ccad702-encryption-config\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038015 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rp5w\" (UniqueName: \"kubernetes.io/projected/05f7659c-85e9-4337-8741-9958f12895d4-kube-api-access-6rp5w\") pod \"router-default-5444994796-fz577\" (UID: \"05f7659c-85e9-4337-8741-9958f12895d4\") " pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038038 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3dbb4222-9650-4663-95bf-bd2fa4f4e678-serving-cert\") pod \"openshift-config-operator-7777fb866f-h6fkg\" (UID: \"3dbb4222-9650-4663-95bf-bd2fa4f4e678\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038072 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp6x8\" (UniqueName: \"kubernetes.io/projected/a9ca675e-e374-467c-aa38-da781b84e3dd-kube-api-access-rp6x8\") pod \"machine-config-operator-74547568cd-cnpz5\" (UID: \"a9ca675e-e374-467c-aa38-da781b84e3dd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038139 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77ws5\" (UniqueName: \"kubernetes.io/projected/a45cf1f5-9679-484d-920c-01e9a0d24587-kube-api-access-77ws5\") pod \"console-operator-58897d9998-2sn7n\" (UID: \"a45cf1f5-9679-484d-920c-01e9a0d24587\") " pod="openshift-console-operator/console-operator-58897d9998-2sn7n" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038165 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/05f7659c-85e9-4337-8741-9958f12895d4-default-certificate\") pod \"router-default-5444994796-fz577\" (UID: \"05f7659c-85e9-4337-8741-9958f12895d4\") " pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038214 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05f7659c-85e9-4337-8741-9958f12895d4-service-ca-bundle\") pod \"router-default-5444994796-fz577\" (UID: \"05f7659c-85e9-4337-8741-9958f12895d4\") " pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038237 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a9ca675e-e374-467c-aa38-da781b84e3dd-auth-proxy-config\") pod \"machine-config-operator-74547568cd-cnpz5\" (UID: \"a9ca675e-e374-467c-aa38-da781b84e3dd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038258 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a45cf1f5-9679-484d-920c-01e9a0d24587-serving-cert\") pod \"console-operator-58897d9998-2sn7n\" (UID: \"a45cf1f5-9679-484d-920c-01e9a0d24587\") " pod="openshift-console-operator/console-operator-58897d9998-2sn7n" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038330 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px568\" (UniqueName: \"kubernetes.io/projected/39dce7bf-1502-40fa-9d9c-41378ccad702-kube-api-access-px568\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038402 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6d2839d4-e624-4d61-9225-894a998e9e9e-console-serving-cert\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038426 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-service-ca\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038470 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/39dce7bf-1502-40fa-9d9c-41378ccad702-node-pullsecrets\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038493 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkks8\" (UniqueName: \"kubernetes.io/projected/6d2839d4-e624-4d61-9225-894a998e9e9e-kube-api-access-vkks8\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038543 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d167fcbe-c833-43f2-ada0-885857d568a9-etcd-ca\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038568 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msqs7\" (UniqueName: \"kubernetes.io/projected/4b90f9f4-fcfa-46b3-80de-b8e335c84e5b-kube-api-access-msqs7\") pod \"downloads-7954f5f757-pvb2q\" (UID: \"4b90f9f4-fcfa-46b3-80de-b8e335c84e5b\") " pod="openshift-console/downloads-7954f5f757-pvb2q" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038608 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6220eaa-3a5d-4c00-8091-fad092c5c77d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gt4k9\" (UID: \"b6220eaa-3a5d-4c00-8091-fad092c5c77d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gt4k9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038632 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/90705934-7171-4811-85d8-8769a47f7fc7-machine-approver-tls\") pod \"machine-approver-56656f9798-l7pnq\" (UID: \"90705934-7171-4811-85d8-8769a47f7fc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038666 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90705934-7171-4811-85d8-8769a47f7fc7-config\") pod \"machine-approver-56656f9798-l7pnq\" (UID: \"90705934-7171-4811-85d8-8769a47f7fc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038688 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-trusted-ca-bundle\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038732 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/39dce7bf-1502-40fa-9d9c-41378ccad702-etcd-serving-ca\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038787 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39dce7bf-1502-40fa-9d9c-41378ccad702-serving-cert\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038815 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3dbb4222-9650-4663-95bf-bd2fa4f4e678-available-featuregates\") pod \"openshift-config-operator-7777fb866f-h6fkg\" (UID: \"3dbb4222-9650-4663-95bf-bd2fa4f4e678\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038849 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr5nb\" (UniqueName: \"kubernetes.io/projected/d167fcbe-c833-43f2-ada0-885857d568a9-kube-api-access-vr5nb\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038872 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a688a250-d68c-4995-b6ff-85daf52874ac-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rl64s\" (UID: \"a688a250-d68c-4995-b6ff-85daf52874ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038893 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/90705934-7171-4811-85d8-8769a47f7fc7-auth-proxy-config\") pod \"machine-approver-56656f9798-l7pnq\" (UID: \"90705934-7171-4811-85d8-8769a47f7fc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038928 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-console-config\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.038962 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039026 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a9ca675e-e374-467c-aa38-da781b84e3dd-proxy-tls\") pod \"machine-config-operator-74547568cd-cnpz5\" (UID: \"a9ca675e-e374-467c-aa38-da781b84e3dd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039057 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d167fcbe-c833-43f2-ada0-885857d568a9-etcd-client\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039080 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06af2589-0701-4b16-a31c-eb39fe1c8b99-config\") pod \"kube-controller-manager-operator-78b949d7b-j25z9\" (UID: \"06af2589-0701-4b16-a31c-eb39fe1c8b99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j25z9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039102 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d167fcbe-c833-43f2-ada0-885857d568a9-config\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039126 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a688a250-d68c-4995-b6ff-85daf52874ac-service-ca-bundle\") pod \"authentication-operator-69f744f599-rl64s\" (UID: \"a688a250-d68c-4995-b6ff-85daf52874ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039148 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d167fcbe-c833-43f2-ada0-885857d568a9-etcd-service-ca\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039170 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9s47\" (UniqueName: \"kubernetes.io/projected/aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721-kube-api-access-b9s47\") pod \"cluster-image-registry-operator-dc59b4c8b-k4pkt\" (UID: \"aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039193 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06af2589-0701-4b16-a31c-eb39fe1c8b99-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-j25z9\" (UID: \"06af2589-0701-4b16-a31c-eb39fe1c8b99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j25z9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039215 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/39dce7bf-1502-40fa-9d9c-41378ccad702-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039239 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/05f7659c-85e9-4337-8741-9958f12895d4-stats-auth\") pod \"router-default-5444994796-fz577\" (UID: \"05f7659c-85e9-4337-8741-9958f12895d4\") " pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039264 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d167fcbe-c833-43f2-ada0-885857d568a9-serving-cert\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039301 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/535884b5-264b-4471-962c-51579eeacab0-trusted-ca\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039327 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/39dce7bf-1502-40fa-9d9c-41378ccad702-audit\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039351 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxwvp\" (UniqueName: \"kubernetes.io/projected/10f43944-25cf-4ae7-8298-82635d81b62e-kube-api-access-nxwvp\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6mbg\" (UID: \"10f43944-25cf-4ae7-8298-82635d81b62e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6mbg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039378 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39dce7bf-1502-40fa-9d9c-41378ccad702-config\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039403 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/39dce7bf-1502-40fa-9d9c-41378ccad702-etcd-client\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039428 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a9ca675e-e374-467c-aa38-da781b84e3dd-images\") pod \"machine-config-operator-74547568cd-cnpz5\" (UID: \"a9ca675e-e374-467c-aa38-da781b84e3dd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039474 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10f43944-25cf-4ae7-8298-82635d81b62e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6mbg\" (UID: \"10f43944-25cf-4ae7-8298-82635d81b62e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6mbg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039498 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/39dce7bf-1502-40fa-9d9c-41378ccad702-audit-dir\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039546 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/dae2df0a-0703-40b5-9040-04b3af885585-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-j8d4t\" (UID: \"dae2df0a-0703-40b5-9040-04b3af885585\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j8d4t" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039595 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/535884b5-264b-4471-962c-51579eeacab0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039618 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a688a250-d68c-4995-b6ff-85daf52874ac-serving-cert\") pod \"authentication-operator-69f744f599-rl64s\" (UID: \"a688a250-d68c-4995-b6ff-85daf52874ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039642 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whxj7\" (UniqueName: \"kubernetes.io/projected/a688a250-d68c-4995-b6ff-85daf52874ac-kube-api-access-whxj7\") pod \"authentication-operator-69f744f599-rl64s\" (UID: \"a688a250-d68c-4995-b6ff-85daf52874ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039664 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6220eaa-3a5d-4c00-8091-fad092c5c77d-config\") pod \"kube-apiserver-operator-766d6c64bb-gt4k9\" (UID: \"b6220eaa-3a5d-4c00-8091-fad092c5c77d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gt4k9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039697 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b6220eaa-3a5d-4c00-8091-fad092c5c77d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gt4k9\" (UID: \"b6220eaa-3a5d-4c00-8091-fad092c5c77d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gt4k9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039723 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sljpv\" (UniqueName: \"kubernetes.io/projected/90705934-7171-4811-85d8-8769a47f7fc7-kube-api-access-sljpv\") pod \"machine-approver-56656f9798-l7pnq\" (UID: \"90705934-7171-4811-85d8-8769a47f7fc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039749 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-oauth-serving-cert\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039774 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/535884b5-264b-4471-962c-51579eeacab0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039796 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10f43944-25cf-4ae7-8298-82635d81b62e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6mbg\" (UID: \"10f43944-25cf-4ae7-8298-82635d81b62e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6mbg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039821 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6d2839d4-e624-4d61-9225-894a998e9e9e-console-oauth-config\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039886 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86x4d\" (UniqueName: \"kubernetes.io/projected/dae2df0a-0703-40b5-9040-04b3af885585-kube-api-access-86x4d\") pod \"cluster-samples-operator-665b6dd947-j8d4t\" (UID: \"dae2df0a-0703-40b5-9040-04b3af885585\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j8d4t" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039911 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a688a250-d68c-4995-b6ff-85daf52874ac-config\") pod \"authentication-operator-69f744f599-rl64s\" (UID: \"a688a250-d68c-4995-b6ff-85daf52874ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039936 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/39dce7bf-1502-40fa-9d9c-41378ccad702-image-import-ca\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039969 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k4pkt\" (UID: \"aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.039992 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-registry-tls\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.040016 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k4pkt\" (UID: \"aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.040041 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a45cf1f5-9679-484d-920c-01e9a0d24587-trusted-ca\") pod \"console-operator-58897d9998-2sn7n\" (UID: \"a45cf1f5-9679-484d-920c-01e9a0d24587\") " pod="openshift-console-operator/console-operator-58897d9998-2sn7n" Dec 03 06:29:59 crc kubenswrapper[4818]: E1203 06:29:59.041674 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:29:59.541656681 +0000 UTC m=+157.233265523 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.094245 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tx2n9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.094426 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58"] Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.138359 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jf227"] Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.141645 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.141783 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b6220eaa-3a5d-4c00-8091-fad092c5c77d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gt4k9\" (UID: \"b6220eaa-3a5d-4c00-8091-fad092c5c77d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gt4k9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.141805 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sljpv\" (UniqueName: \"kubernetes.io/projected/90705934-7171-4811-85d8-8769a47f7fc7-kube-api-access-sljpv\") pod \"machine-approver-56656f9798-l7pnq\" (UID: \"90705934-7171-4811-85d8-8769a47f7fc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.141843 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq877\" (UniqueName: \"kubernetes.io/projected/a3eedabc-4a51-4068-a9da-c8a867cb3219-kube-api-access-vq877\") pod \"package-server-manager-789f6589d5-6glnm\" (UID: \"a3eedabc-4a51-4068-a9da-c8a867cb3219\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6glnm" Dec 03 06:29:59 crc kubenswrapper[4818]: E1203 06:29:59.141892 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:29:59.64186367 +0000 UTC m=+157.333472502 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.141934 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/535884b5-264b-4471-962c-51579eeacab0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.141980 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a63ab77-ab0e-4f43-9d21-2ec684236ce9-config-volume\") pod \"dns-default-79bwt\" (UID: \"3a63ab77-ab0e-4f43-9d21-2ec684236ce9\") " pod="openshift-dns/dns-default-79bwt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142017 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-bv9gv\" (UID: \"1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142044 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b14853e5-2388-4dba-a3ca-dd9453ea3452-mountpoint-dir\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142068 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtgcj\" (UniqueName: \"kubernetes.io/projected/3a63ab77-ab0e-4f43-9d21-2ec684236ce9-kube-api-access-dtgcj\") pod \"dns-default-79bwt\" (UID: \"3a63ab77-ab0e-4f43-9d21-2ec684236ce9\") " pod="openshift-dns/dns-default-79bwt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142097 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a688a250-d68c-4995-b6ff-85daf52874ac-config\") pod \"authentication-operator-69f744f599-rl64s\" (UID: \"a688a250-d68c-4995-b6ff-85daf52874ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142121 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/39dce7bf-1502-40fa-9d9c-41378ccad702-image-import-ca\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142152 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k4pkt\" (UID: \"aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142177 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fcz6\" (UniqueName: \"kubernetes.io/projected/503c8137-bff7-476c-90f1-6c0d8112cc2e-kube-api-access-2fcz6\") pod \"olm-operator-6b444d44fb-dr7rf\" (UID: \"503c8137-bff7-476c-90f1-6c0d8112cc2e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142204 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4cll\" (UniqueName: \"kubernetes.io/projected/0aef8214-0bed-45cc-b3cd-77fcfbecb7aa-kube-api-access-p4cll\") pod \"catalog-operator-68c6474976-jr26c\" (UID: \"0aef8214-0bed-45cc-b3cd-77fcfbecb7aa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142229 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c8c049a4-28db-4ac4-807c-637c3837235a-node-bootstrap-token\") pod \"machine-config-server-l6nvl\" (UID: \"c8c049a4-28db-4ac4-807c-637c3837235a\") " pod="openshift-machine-config-operator/machine-config-server-l6nvl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142256 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/535884b5-264b-4471-962c-51579eeacab0-registry-certificates\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142278 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a45cf1f5-9679-484d-920c-01e9a0d24587-config\") pod \"console-operator-58897d9998-2sn7n\" (UID: \"a45cf1f5-9679-484d-920c-01e9a0d24587\") " pod="openshift-console-operator/console-operator-58897d9998-2sn7n" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142302 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-bound-sa-token\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142324 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/05f7659c-85e9-4337-8741-9958f12895d4-metrics-certs\") pod \"router-default-5444994796-fz577\" (UID: \"05f7659c-85e9-4337-8741-9958f12895d4\") " pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142348 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c633e2b-4169-4277-a64b-85c249869b0c-webhook-cert\") pod \"packageserver-d55dfcdfc-f55bl\" (UID: \"0c633e2b-4169-4277-a64b-85c249869b0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142374 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a2888eb-7da9-4ef2-841b-22f170e66a54-serving-cert\") pod \"service-ca-operator-777779d784-phwq8\" (UID: \"9a2888eb-7da9-4ef2-841b-22f170e66a54\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-phwq8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142397 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgdgs\" (UniqueName: \"kubernetes.io/projected/9a2888eb-7da9-4ef2-841b-22f170e66a54-kube-api-access-qgdgs\") pod \"service-ca-operator-777779d784-phwq8\" (UID: \"9a2888eb-7da9-4ef2-841b-22f170e66a54\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-phwq8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142423 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8935f198-82ab-463c-96d5-300a23d613e6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s8mcw\" (UID: \"8935f198-82ab-463c-96d5-300a23d613e6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s8mcw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142447 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77ws5\" (UniqueName: \"kubernetes.io/projected/a45cf1f5-9679-484d-920c-01e9a0d24587-kube-api-access-77ws5\") pod \"console-operator-58897d9998-2sn7n\" (UID: \"a45cf1f5-9679-484d-920c-01e9a0d24587\") " pod="openshift-console-operator/console-operator-58897d9998-2sn7n" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142471 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/05f7659c-85e9-4337-8741-9958f12895d4-default-certificate\") pod \"router-default-5444994796-fz577\" (UID: \"05f7659c-85e9-4337-8741-9958f12895d4\") " pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142493 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3dbb4222-9650-4663-95bf-bd2fa4f4e678-serving-cert\") pod \"openshift-config-operator-7777fb866f-h6fkg\" (UID: \"3dbb4222-9650-4663-95bf-bd2fa4f4e678\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142517 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a3eedabc-4a51-4068-a9da-c8a867cb3219-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-6glnm\" (UID: \"a3eedabc-4a51-4068-a9da-c8a867cb3219\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6glnm" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142557 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0c633e2b-4169-4277-a64b-85c249869b0c-tmpfs\") pod \"packageserver-d55dfcdfc-f55bl\" (UID: \"0c633e2b-4169-4277-a64b-85c249869b0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142584 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a9ca675e-e374-467c-aa38-da781b84e3dd-auth-proxy-config\") pod \"machine-config-operator-74547568cd-cnpz5\" (UID: \"a9ca675e-e374-467c-aa38-da781b84e3dd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142606 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a45cf1f5-9679-484d-920c-01e9a0d24587-serving-cert\") pod \"console-operator-58897d9998-2sn7n\" (UID: \"a45cf1f5-9679-484d-920c-01e9a0d24587\") " pod="openshift-console-operator/console-operator-58897d9998-2sn7n" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142630 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b14853e5-2388-4dba-a3ca-dd9453ea3452-csi-data-dir\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142656 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-service-ca\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142679 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/39dce7bf-1502-40fa-9d9c-41378ccad702-node-pullsecrets\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142703 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7k47\" (UniqueName: \"kubernetes.io/projected/b14853e5-2388-4dba-a3ca-dd9453ea3452-kube-api-access-g7k47\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142728 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90705934-7171-4811-85d8-8769a47f7fc7-config\") pod \"machine-approver-56656f9798-l7pnq\" (UID: \"90705934-7171-4811-85d8-8769a47f7fc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142752 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0ff5cea8-7b98-4a15-87cb-f07fb9cfabea-signing-cabundle\") pod \"service-ca-9c57cc56f-596cp\" (UID: \"0ff5cea8-7b98-4a15-87cb-f07fb9cfabea\") " pod="openshift-service-ca/service-ca-9c57cc56f-596cp" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142787 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d167fcbe-c833-43f2-ada0-885857d568a9-etcd-ca\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142816 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6220eaa-3a5d-4c00-8091-fad092c5c77d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gt4k9\" (UID: \"b6220eaa-3a5d-4c00-8091-fad092c5c77d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gt4k9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142855 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/90705934-7171-4811-85d8-8769a47f7fc7-machine-approver-tls\") pod \"machine-approver-56656f9798-l7pnq\" (UID: \"90705934-7171-4811-85d8-8769a47f7fc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142893 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-trusted-ca-bundle\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142947 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b14853e5-2388-4dba-a3ca-dd9453ea3452-registration-dir\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.142979 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a688a250-d68c-4995-b6ff-85daf52874ac-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rl64s\" (UID: \"a688a250-d68c-4995-b6ff-85daf52874ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.143043 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a688a250-d68c-4995-b6ff-85daf52874ac-config\") pod \"authentication-operator-69f744f599-rl64s\" (UID: \"a688a250-d68c-4995-b6ff-85daf52874ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.143167 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39dce7bf-1502-40fa-9d9c-41378ccad702-serving-cert\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.143218 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3dbb4222-9650-4663-95bf-bd2fa4f4e678-available-featuregates\") pod \"openshift-config-operator-7777fb866f-h6fkg\" (UID: \"3dbb4222-9650-4663-95bf-bd2fa4f4e678\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.143240 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6-proxy-tls\") pod \"machine-config-controller-84d6567774-bv9gv\" (UID: \"1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.143304 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-console-config\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.143324 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8935f198-82ab-463c-96d5-300a23d613e6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s8mcw\" (UID: \"8935f198-82ab-463c-96d5-300a23d613e6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s8mcw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.143340 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c313309-b494-4209-a248-cb526aa88f36-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dh7z8\" (UID: \"2c313309-b494-4209-a248-cb526aa88f36\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dh7z8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.143472 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.143493 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/faad9125-f494-49ae-ba77-941e6abe67b9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ltb69\" (UID: \"faad9125-f494-49ae-ba77-941e6abe67b9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ltb69" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.143522 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9d0ddd49-ca6f-44aa-a3aa-5a7717845156-metrics-tls\") pod \"dns-operator-744455d44c-jvbsr\" (UID: \"9d0ddd49-ca6f-44aa-a3aa-5a7717845156\") " pod="openshift-dns-operator/dns-operator-744455d44c-jvbsr" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.143540 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b14853e5-2388-4dba-a3ca-dd9453ea3452-socket-dir\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.143558 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d167fcbe-c833-43f2-ada0-885857d568a9-config\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.143762 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a45cf1f5-9679-484d-920c-01e9a0d24587-config\") pod \"console-operator-58897d9998-2sn7n\" (UID: \"a45cf1f5-9679-484d-920c-01e9a0d24587\") " pod="openshift-console-operator/console-operator-58897d9998-2sn7n" Dec 03 06:29:59 crc kubenswrapper[4818]: E1203 06:29:59.143899 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:29:59.643887171 +0000 UTC m=+157.335496023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.143947 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/39dce7bf-1502-40fa-9d9c-41378ccad702-node-pullsecrets\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.144359 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/535884b5-264b-4471-962c-51579eeacab0-registry-certificates\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.144504 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-trusted-ca-bundle\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.144970 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a688a250-d68c-4995-b6ff-85daf52874ac-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rl64s\" (UID: \"a688a250-d68c-4995-b6ff-85daf52874ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.145288 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-console-config\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.145355 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/53a413a7-34f8-4fb3-893b-c4fa9711017a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4dh8d\" (UID: \"53a413a7-34f8-4fb3-893b-c4fa9711017a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.145388 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-config-volume\") pod \"collect-profiles-29412375-f596r\" (UID: \"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.145461 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d167fcbe-c833-43f2-ada0-885857d568a9-serving-cert\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.145487 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/05f7659c-85e9-4337-8741-9958f12895d4-stats-auth\") pod \"router-default-5444994796-fz577\" (UID: \"05f7659c-85e9-4337-8741-9958f12895d4\") " pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.145512 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8935f198-82ab-463c-96d5-300a23d613e6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s8mcw\" (UID: \"8935f198-82ab-463c-96d5-300a23d613e6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s8mcw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.145538 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdv8l\" (UniqueName: \"kubernetes.io/projected/2bf3947e-3c18-4ef1-aca9-98419c558d17-kube-api-access-gdv8l\") pod \"ingress-operator-5b745b69d9-8jzsz\" (UID: \"2bf3947e-3c18-4ef1-aca9-98419c558d17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.145554 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a9ca675e-e374-467c-aa38-da781b84e3dd-auth-proxy-config\") pod \"machine-config-operator-74547568cd-cnpz5\" (UID: \"a9ca675e-e374-467c-aa38-da781b84e3dd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.145599 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/535884b5-264b-4471-962c-51579eeacab0-trusted-ca\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.145625 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbvwx\" (UniqueName: \"kubernetes.io/projected/faad9125-f494-49ae-ba77-941e6abe67b9-kube-api-access-pbvwx\") pod \"control-plane-machine-set-operator-78cbb6b69f-ltb69\" (UID: \"faad9125-f494-49ae-ba77-941e6abe67b9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ltb69" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.145651 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/503c8137-bff7-476c-90f1-6c0d8112cc2e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-dr7rf\" (UID: \"503c8137-bff7-476c-90f1-6c0d8112cc2e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.145692 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10f43944-25cf-4ae7-8298-82635d81b62e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6mbg\" (UID: \"10f43944-25cf-4ae7-8298-82635d81b62e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6mbg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.145718 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxwvp\" (UniqueName: \"kubernetes.io/projected/10f43944-25cf-4ae7-8298-82635d81b62e-kube-api-access-nxwvp\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6mbg\" (UID: \"10f43944-25cf-4ae7-8298-82635d81b62e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6mbg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.145747 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39dce7bf-1502-40fa-9d9c-41378ccad702-config\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.145770 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/39dce7bf-1502-40fa-9d9c-41378ccad702-etcd-client\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.146319 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/39dce7bf-1502-40fa-9d9c-41378ccad702-image-import-ca\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.146469 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90705934-7171-4811-85d8-8769a47f7fc7-config\") pod \"machine-approver-56656f9798-l7pnq\" (UID: \"90705934-7171-4811-85d8-8769a47f7fc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.146566 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d167fcbe-c833-43f2-ada0-885857d568a9-etcd-ca\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.147388 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-service-ca\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.148032 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d167fcbe-c833-43f2-ada0-885857d568a9-config\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.148630 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39dce7bf-1502-40fa-9d9c-41378ccad702-config\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.149021 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3dbb4222-9650-4663-95bf-bd2fa4f4e678-available-featuregates\") pod \"openshift-config-operator-7777fb866f-h6fkg\" (UID: \"3dbb4222-9650-4663-95bf-bd2fa4f4e678\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.149434 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/535884b5-264b-4471-962c-51579eeacab0-trusted-ca\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.149752 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k4pkt\" (UID: \"aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.149784 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/05f7659c-85e9-4337-8741-9958f12895d4-default-certificate\") pod \"router-default-5444994796-fz577\" (UID: \"05f7659c-85e9-4337-8741-9958f12895d4\") " pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.149952 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a688a250-d68c-4995-b6ff-85daf52874ac-serving-cert\") pod \"authentication-operator-69f744f599-rl64s\" (UID: \"a688a250-d68c-4995-b6ff-85daf52874ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.150021 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whxj7\" (UniqueName: \"kubernetes.io/projected/a688a250-d68c-4995-b6ff-85daf52874ac-kube-api-access-whxj7\") pod \"authentication-operator-69f744f599-rl64s\" (UID: \"a688a250-d68c-4995-b6ff-85daf52874ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.150049 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6220eaa-3a5d-4c00-8091-fad092c5c77d-config\") pod \"kube-apiserver-operator-766d6c64bb-gt4k9\" (UID: \"b6220eaa-3a5d-4c00-8091-fad092c5c77d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gt4k9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.150072 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-oauth-serving-cert\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.151027 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10f43944-25cf-4ae7-8298-82635d81b62e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6mbg\" (UID: \"10f43944-25cf-4ae7-8298-82635d81b62e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6mbg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.151095 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlm4m\" (UniqueName: \"kubernetes.io/projected/53a413a7-34f8-4fb3-893b-c4fa9711017a-kube-api-access-vlm4m\") pod \"marketplace-operator-79b997595-4dh8d\" (UID: \"53a413a7-34f8-4fb3-893b-c4fa9711017a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.151134 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6d2839d4-e624-4d61-9225-894a998e9e9e-console-oauth-config\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.151199 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86x4d\" (UniqueName: \"kubernetes.io/projected/dae2df0a-0703-40b5-9040-04b3af885585-kube-api-access-86x4d\") pod \"cluster-samples-operator-665b6dd947-j8d4t\" (UID: \"dae2df0a-0703-40b5-9040-04b3af885585\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j8d4t" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.151229 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2bf3947e-3c18-4ef1-aca9-98419c558d17-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8jzsz\" (UID: \"2bf3947e-3c18-4ef1-aca9-98419c558d17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.151270 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k4pkt\" (UID: \"aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.151351 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-registry-tls\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.151378 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a45cf1f5-9679-484d-920c-01e9a0d24587-trusted-ca\") pod \"console-operator-58897d9998-2sn7n\" (UID: \"a45cf1f5-9679-484d-920c-01e9a0d24587\") " pod="openshift-console-operator/console-operator-58897d9998-2sn7n" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.151404 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0ff5cea8-7b98-4a15-87cb-f07fb9cfabea-signing-key\") pod \"service-ca-9c57cc56f-596cp\" (UID: \"0ff5cea8-7b98-4a15-87cb-f07fb9cfabea\") " pod="openshift-service-ca/service-ca-9c57cc56f-596cp" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.151426 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0aef8214-0bed-45cc-b3cd-77fcfbecb7aa-profile-collector-cert\") pod \"catalog-operator-68c6474976-jr26c\" (UID: \"0aef8214-0bed-45cc-b3cd-77fcfbecb7aa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.151424 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b6220eaa-3a5d-4c00-8091-fad092c5c77d-config\") pod \"kube-apiserver-operator-766d6c64bb-gt4k9\" (UID: \"b6220eaa-3a5d-4c00-8091-fad092c5c77d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gt4k9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.151447 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c8c049a4-28db-4ac4-807c-637c3837235a-certs\") pod \"machine-config-server-l6nvl\" (UID: \"c8c049a4-28db-4ac4-807c-637c3837235a\") " pod="openshift-machine-config-operator/machine-config-server-l6nvl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.152040 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-oauth-serving-cert\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.152195 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/06af2589-0701-4b16-a31c-eb39fe1c8b99-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-j25z9\" (UID: \"06af2589-0701-4b16-a31c-eb39fe1c8b99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j25z9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.152228 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9nsd\" (UniqueName: \"kubernetes.io/projected/c8c049a4-28db-4ac4-807c-637c3837235a-kube-api-access-n9nsd\") pod \"machine-config-server-l6nvl\" (UID: \"c8c049a4-28db-4ac4-807c-637c3837235a\") " pod="openshift-machine-config-operator/machine-config-server-l6nvl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.152295 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnlc6\" (UniqueName: \"kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-kube-api-access-gnlc6\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.152323 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k4pkt\" (UID: \"aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.152346 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/39dce7bf-1502-40fa-9d9c-41378ccad702-encryption-config\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.152370 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnmvg\" (UniqueName: \"kubernetes.io/projected/3dbb4222-9650-4663-95bf-bd2fa4f4e678-kube-api-access-cnmvg\") pod \"openshift-config-operator-7777fb866f-h6fkg\" (UID: \"3dbb4222-9650-4663-95bf-bd2fa4f4e678\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.152397 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwrzj\" (UniqueName: \"kubernetes.io/projected/0ff5cea8-7b98-4a15-87cb-f07fb9cfabea-kube-api-access-nwrzj\") pod \"service-ca-9c57cc56f-596cp\" (UID: \"0ff5cea8-7b98-4a15-87cb-f07fb9cfabea\") " pod="openshift-service-ca/service-ca-9c57cc56f-596cp" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.152420 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp6x8\" (UniqueName: \"kubernetes.io/projected/a9ca675e-e374-467c-aa38-da781b84e3dd-kube-api-access-rp6x8\") pod \"machine-config-operator-74547568cd-cnpz5\" (UID: \"a9ca675e-e374-467c-aa38-da781b84e3dd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.152511 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rp5w\" (UniqueName: \"kubernetes.io/projected/05f7659c-85e9-4337-8741-9958f12895d4-kube-api-access-6rp5w\") pod \"router-default-5444994796-fz577\" (UID: \"05f7659c-85e9-4337-8741-9958f12895d4\") " pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.152533 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05f7659c-85e9-4337-8741-9958f12895d4-service-ca-bundle\") pod \"router-default-5444994796-fz577\" (UID: \"05f7659c-85e9-4337-8741-9958f12895d4\") " pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.152557 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q4s7\" (UniqueName: \"kubernetes.io/projected/8b9018a9-c76c-4471-8b27-06c1cbc065eb-kube-api-access-4q4s7\") pod \"multus-admission-controller-857f4d67dd-s8vm2\" (UID: \"8b9018a9-c76c-4471-8b27-06c1cbc065eb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s8vm2" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.152582 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a2888eb-7da9-4ef2-841b-22f170e66a54-config\") pod \"service-ca-operator-777779d784-phwq8\" (UID: \"9a2888eb-7da9-4ef2-841b-22f170e66a54\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-phwq8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.152606 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px568\" (UniqueName: \"kubernetes.io/projected/39dce7bf-1502-40fa-9d9c-41378ccad702-kube-api-access-px568\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.152626 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c633e2b-4169-4277-a64b-85c249869b0c-apiservice-cert\") pod \"packageserver-d55dfcdfc-f55bl\" (UID: \"0c633e2b-4169-4277-a64b-85c249869b0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.152648 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5cdd\" (UniqueName: \"kubernetes.io/projected/0c633e2b-4169-4277-a64b-85c249869b0c-kube-api-access-r5cdd\") pod \"packageserver-d55dfcdfc-f55bl\" (UID: \"0c633e2b-4169-4277-a64b-85c249869b0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.152669 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6d2839d4-e624-4d61-9225-894a998e9e9e-console-serving-cert\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.153428 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a45cf1f5-9679-484d-920c-01e9a0d24587-trusted-ca\") pod \"console-operator-58897d9998-2sn7n\" (UID: \"a45cf1f5-9679-484d-920c-01e9a0d24587\") " pod="openshift-console-operator/console-operator-58897d9998-2sn7n" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.153683 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkks8\" (UniqueName: \"kubernetes.io/projected/6d2839d4-e624-4d61-9225-894a998e9e9e-kube-api-access-vkks8\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.153704 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a688a250-d68c-4995-b6ff-85daf52874ac-serving-cert\") pod \"authentication-operator-69f744f599-rl64s\" (UID: \"a688a250-d68c-4995-b6ff-85daf52874ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.153873 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05f7659c-85e9-4337-8741-9958f12895d4-service-ca-bundle\") pod \"router-default-5444994796-fz577\" (UID: \"05f7659c-85e9-4337-8741-9958f12895d4\") " pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.154055 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msqs7\" (UniqueName: \"kubernetes.io/projected/4b90f9f4-fcfa-46b3-80de-b8e335c84e5b-kube-api-access-msqs7\") pod \"downloads-7954f5f757-pvb2q\" (UID: \"4b90f9f4-fcfa-46b3-80de-b8e335c84e5b\") " pod="openshift-console/downloads-7954f5f757-pvb2q" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.154115 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-secret-volume\") pod \"collect-profiles-29412375-f596r\" (UID: \"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.154145 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2bf3947e-3c18-4ef1-aca9-98419c558d17-trusted-ca\") pod \"ingress-operator-5b745b69d9-8jzsz\" (UID: \"2bf3947e-3c18-4ef1-aca9-98419c558d17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.154194 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c313309-b494-4209-a248-cb526aa88f36-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dh7z8\" (UID: \"2c313309-b494-4209-a248-cb526aa88f36\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dh7z8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.154303 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/39dce7bf-1502-40fa-9d9c-41378ccad702-etcd-serving-ca\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.154326 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10f43944-25cf-4ae7-8298-82635d81b62e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6mbg\" (UID: \"10f43944-25cf-4ae7-8298-82635d81b62e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6mbg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.154377 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxjhq\" (UniqueName: \"kubernetes.io/projected/1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6-kube-api-access-xxjhq\") pod \"machine-config-controller-84d6567774-bv9gv\" (UID: \"1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.154473 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53a413a7-34f8-4fb3-893b-c4fa9711017a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4dh8d\" (UID: \"53a413a7-34f8-4fb3-893b-c4fa9711017a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.154595 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flzcx\" (UniqueName: \"kubernetes.io/projected/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-kube-api-access-flzcx\") pod \"collect-profiles-29412375-f596r\" (UID: \"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.154770 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/39dce7bf-1502-40fa-9d9c-41378ccad702-etcd-serving-ca\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.154908 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vr5nb\" (UniqueName: \"kubernetes.io/projected/d167fcbe-c833-43f2-ada0-885857d568a9-kube-api-access-vr5nb\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.154936 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/90705934-7171-4811-85d8-8769a47f7fc7-auth-proxy-config\") pod \"machine-approver-56656f9798-l7pnq\" (UID: \"90705934-7171-4811-85d8-8769a47f7fc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.154957 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/503c8137-bff7-476c-90f1-6c0d8112cc2e-srv-cert\") pod \"olm-operator-6b444d44fb-dr7rf\" (UID: \"503c8137-bff7-476c-90f1-6c0d8112cc2e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.155140 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xngvs\" (UniqueName: \"kubernetes.io/projected/a0dc41b5-d81d-4a47-aaad-8a9d8216e513-kube-api-access-xngvs\") pod \"ingress-canary-nmqf4\" (UID: \"a0dc41b5-d81d-4a47-aaad-8a9d8216e513\") " pod="openshift-ingress-canary/ingress-canary-nmqf4" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.155220 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a9ca675e-e374-467c-aa38-da781b84e3dd-proxy-tls\") pod \"machine-config-operator-74547568cd-cnpz5\" (UID: \"a9ca675e-e374-467c-aa38-da781b84e3dd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156015 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-registry-tls\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156022 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/90705934-7171-4811-85d8-8769a47f7fc7-auth-proxy-config\") pod \"machine-approver-56656f9798-l7pnq\" (UID: \"90705934-7171-4811-85d8-8769a47f7fc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156069 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwh6f\" (UniqueName: \"kubernetes.io/projected/2c313309-b494-4209-a248-cb526aa88f36-kube-api-access-qwh6f\") pod \"kube-storage-version-migrator-operator-b67b599dd-dh7z8\" (UID: \"2c313309-b494-4209-a248-cb526aa88f36\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dh7z8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156105 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d167fcbe-c833-43f2-ada0-885857d568a9-etcd-client\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156131 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06af2589-0701-4b16-a31c-eb39fe1c8b99-config\") pod \"kube-controller-manager-operator-78b949d7b-j25z9\" (UID: \"06af2589-0701-4b16-a31c-eb39fe1c8b99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j25z9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156160 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a688a250-d68c-4995-b6ff-85daf52874ac-service-ca-bundle\") pod \"authentication-operator-69f744f599-rl64s\" (UID: \"a688a250-d68c-4995-b6ff-85daf52874ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156184 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d167fcbe-c833-43f2-ada0-885857d568a9-etcd-service-ca\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156207 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9s47\" (UniqueName: \"kubernetes.io/projected/aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721-kube-api-access-b9s47\") pod \"cluster-image-registry-operator-dc59b4c8b-k4pkt\" (UID: \"aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156230 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06af2589-0701-4b16-a31c-eb39fe1c8b99-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-j25z9\" (UID: \"06af2589-0701-4b16-a31c-eb39fe1c8b99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j25z9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156254 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp4sk\" (UniqueName: \"kubernetes.io/projected/9d0ddd49-ca6f-44aa-a3aa-5a7717845156-kube-api-access-cp4sk\") pod \"dns-operator-744455d44c-jvbsr\" (UID: \"9d0ddd49-ca6f-44aa-a3aa-5a7717845156\") " pod="openshift-dns-operator/dns-operator-744455d44c-jvbsr" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156277 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2bf3947e-3c18-4ef1-aca9-98419c558d17-metrics-tls\") pod \"ingress-operator-5b745b69d9-8jzsz\" (UID: \"2bf3947e-3c18-4ef1-aca9-98419c558d17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156304 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3a63ab77-ab0e-4f43-9d21-2ec684236ce9-metrics-tls\") pod \"dns-default-79bwt\" (UID: \"3a63ab77-ab0e-4f43-9d21-2ec684236ce9\") " pod="openshift-dns/dns-default-79bwt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156330 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/39dce7bf-1502-40fa-9d9c-41378ccad702-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156354 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b9c2\" (UniqueName: \"kubernetes.io/projected/3deb8606-51f3-4a91-bcfd-0533f315e829-kube-api-access-8b9c2\") pod \"migrator-59844c95c7-8x7mv\" (UID: \"3deb8606-51f3-4a91-bcfd-0533f315e829\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8x7mv" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156377 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b14853e5-2388-4dba-a3ca-dd9453ea3452-plugins-dir\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156395 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0dc41b5-d81d-4a47-aaad-8a9d8216e513-cert\") pod \"ingress-canary-nmqf4\" (UID: \"a0dc41b5-d81d-4a47-aaad-8a9d8216e513\") " pod="openshift-ingress-canary/ingress-canary-nmqf4" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156415 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0aef8214-0bed-45cc-b3cd-77fcfbecb7aa-srv-cert\") pod \"catalog-operator-68c6474976-jr26c\" (UID: \"0aef8214-0bed-45cc-b3cd-77fcfbecb7aa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156453 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/39dce7bf-1502-40fa-9d9c-41378ccad702-audit\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156477 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a9ca675e-e374-467c-aa38-da781b84e3dd-images\") pod \"machine-config-operator-74547568cd-cnpz5\" (UID: \"a9ca675e-e374-467c-aa38-da781b84e3dd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156500 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8b9018a9-c76c-4471-8b27-06c1cbc065eb-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-s8vm2\" (UID: \"8b9018a9-c76c-4471-8b27-06c1cbc065eb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s8vm2" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156524 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/39dce7bf-1502-40fa-9d9c-41378ccad702-audit-dir\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156546 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/dae2df0a-0703-40b5-9040-04b3af885585-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-j8d4t\" (UID: \"dae2df0a-0703-40b5-9040-04b3af885585\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j8d4t" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.156569 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/535884b5-264b-4471-962c-51579eeacab0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.157037 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/535884b5-264b-4471-962c-51579eeacab0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.157648 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/39dce7bf-1502-40fa-9d9c-41378ccad702-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.157915 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6d2839d4-e624-4d61-9225-894a998e9e9e-console-serving-cert\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.157988 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/39dce7bf-1502-40fa-9d9c-41378ccad702-audit\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.158152 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/39dce7bf-1502-40fa-9d9c-41378ccad702-encryption-config\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.158243 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/39dce7bf-1502-40fa-9d9c-41378ccad702-audit-dir\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.158774 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a9ca675e-e374-467c-aa38-da781b84e3dd-images\") pod \"machine-config-operator-74547568cd-cnpz5\" (UID: \"a9ca675e-e374-467c-aa38-da781b84e3dd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.159143 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k4pkt\" (UID: \"aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.159207 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d167fcbe-c833-43f2-ada0-885857d568a9-serving-cert\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.159321 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a688a250-d68c-4995-b6ff-85daf52874ac-service-ca-bundle\") pod \"authentication-operator-69f744f599-rl64s\" (UID: \"a688a250-d68c-4995-b6ff-85daf52874ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.159700 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6d2839d4-e624-4d61-9225-894a998e9e9e-console-oauth-config\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.159707 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/05f7659c-85e9-4337-8741-9958f12895d4-stats-auth\") pod \"router-default-5444994796-fz577\" (UID: \"05f7659c-85e9-4337-8741-9958f12895d4\") " pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.159844 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d167fcbe-c833-43f2-ada0-885857d568a9-etcd-service-ca\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.159863 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a9ca675e-e374-467c-aa38-da781b84e3dd-proxy-tls\") pod \"machine-config-operator-74547568cd-cnpz5\" (UID: \"a9ca675e-e374-467c-aa38-da781b84e3dd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.160372 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/39dce7bf-1502-40fa-9d9c-41378ccad702-etcd-client\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.161435 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a45cf1f5-9679-484d-920c-01e9a0d24587-serving-cert\") pod \"console-operator-58897d9998-2sn7n\" (UID: \"a45cf1f5-9679-484d-920c-01e9a0d24587\") " pod="openshift-console-operator/console-operator-58897d9998-2sn7n" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.162944 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/535884b5-264b-4471-962c-51579eeacab0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.164134 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3dbb4222-9650-4663-95bf-bd2fa4f4e678-serving-cert\") pod \"openshift-config-operator-7777fb866f-h6fkg\" (UID: \"3dbb4222-9650-4663-95bf-bd2fa4f4e678\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.164470 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/05f7659c-85e9-4337-8741-9958f12895d4-metrics-certs\") pod \"router-default-5444994796-fz577\" (UID: \"05f7659c-85e9-4337-8741-9958f12895d4\") " pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.164672 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/90705934-7171-4811-85d8-8769a47f7fc7-machine-approver-tls\") pod \"machine-approver-56656f9798-l7pnq\" (UID: \"90705934-7171-4811-85d8-8769a47f7fc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.164755 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d167fcbe-c833-43f2-ada0-885857d568a9-etcd-client\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.165239 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/dae2df0a-0703-40b5-9040-04b3af885585-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-j8d4t\" (UID: \"dae2df0a-0703-40b5-9040-04b3af885585\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j8d4t" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.165989 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39dce7bf-1502-40fa-9d9c-41378ccad702-serving-cert\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.166626 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06af2589-0701-4b16-a31c-eb39fe1c8b99-config\") pod \"kube-controller-manager-operator-78b949d7b-j25z9\" (UID: \"06af2589-0701-4b16-a31c-eb39fe1c8b99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j25z9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.167843 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06af2589-0701-4b16-a31c-eb39fe1c8b99-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-j25z9\" (UID: \"06af2589-0701-4b16-a31c-eb39fe1c8b99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j25z9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.173123 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10f43944-25cf-4ae7-8298-82635d81b62e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6mbg\" (UID: \"10f43944-25cf-4ae7-8298-82635d81b62e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6mbg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.173267 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b6220eaa-3a5d-4c00-8091-fad092c5c77d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gt4k9\" (UID: \"b6220eaa-3a5d-4c00-8091-fad092c5c77d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gt4k9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.174912 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b6220eaa-3a5d-4c00-8091-fad092c5c77d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gt4k9\" (UID: \"b6220eaa-3a5d-4c00-8091-fad092c5c77d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gt4k9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.181888 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z"] Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.191479 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.207709 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sljpv\" (UniqueName: \"kubernetes.io/projected/90705934-7171-4811-85d8-8769a47f7fc7-kube-api-access-sljpv\") pod \"machine-approver-56656f9798-l7pnq\" (UID: \"90705934-7171-4811-85d8-8769a47f7fc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.236898 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-bound-sa-token\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.257451 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.257688 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp4sk\" (UniqueName: \"kubernetes.io/projected/9d0ddd49-ca6f-44aa-a3aa-5a7717845156-kube-api-access-cp4sk\") pod \"dns-operator-744455d44c-jvbsr\" (UID: \"9d0ddd49-ca6f-44aa-a3aa-5a7717845156\") " pod="openshift-dns-operator/dns-operator-744455d44c-jvbsr" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.257906 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b9c2\" (UniqueName: \"kubernetes.io/projected/3deb8606-51f3-4a91-bcfd-0533f315e829-kube-api-access-8b9c2\") pod \"migrator-59844c95c7-8x7mv\" (UID: \"3deb8606-51f3-4a91-bcfd-0533f315e829\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8x7mv" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258238 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b14853e5-2388-4dba-a3ca-dd9453ea3452-plugins-dir\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.257999 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b14853e5-2388-4dba-a3ca-dd9453ea3452-plugins-dir\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258334 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2bf3947e-3c18-4ef1-aca9-98419c558d17-metrics-tls\") pod \"ingress-operator-5b745b69d9-8jzsz\" (UID: \"2bf3947e-3c18-4ef1-aca9-98419c558d17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258357 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3a63ab77-ab0e-4f43-9d21-2ec684236ce9-metrics-tls\") pod \"dns-default-79bwt\" (UID: \"3a63ab77-ab0e-4f43-9d21-2ec684236ce9\") " pod="openshift-dns/dns-default-79bwt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258378 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0dc41b5-d81d-4a47-aaad-8a9d8216e513-cert\") pod \"ingress-canary-nmqf4\" (UID: \"a0dc41b5-d81d-4a47-aaad-8a9d8216e513\") " pod="openshift-ingress-canary/ingress-canary-nmqf4" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258402 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0aef8214-0bed-45cc-b3cd-77fcfbecb7aa-srv-cert\") pod \"catalog-operator-68c6474976-jr26c\" (UID: \"0aef8214-0bed-45cc-b3cd-77fcfbecb7aa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258424 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8b9018a9-c76c-4471-8b27-06c1cbc065eb-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-s8vm2\" (UID: \"8b9018a9-c76c-4471-8b27-06c1cbc065eb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s8vm2" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258449 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq877\" (UniqueName: \"kubernetes.io/projected/a3eedabc-4a51-4068-a9da-c8a867cb3219-kube-api-access-vq877\") pod \"package-server-manager-789f6589d5-6glnm\" (UID: \"a3eedabc-4a51-4068-a9da-c8a867cb3219\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6glnm" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258475 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a63ab77-ab0e-4f43-9d21-2ec684236ce9-config-volume\") pod \"dns-default-79bwt\" (UID: \"3a63ab77-ab0e-4f43-9d21-2ec684236ce9\") " pod="openshift-dns/dns-default-79bwt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258498 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-bv9gv\" (UID: \"1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258517 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b14853e5-2388-4dba-a3ca-dd9453ea3452-mountpoint-dir\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258539 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtgcj\" (UniqueName: \"kubernetes.io/projected/3a63ab77-ab0e-4f43-9d21-2ec684236ce9-kube-api-access-dtgcj\") pod \"dns-default-79bwt\" (UID: \"3a63ab77-ab0e-4f43-9d21-2ec684236ce9\") " pod="openshift-dns/dns-default-79bwt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258564 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fcz6\" (UniqueName: \"kubernetes.io/projected/503c8137-bff7-476c-90f1-6c0d8112cc2e-kube-api-access-2fcz6\") pod \"olm-operator-6b444d44fb-dr7rf\" (UID: \"503c8137-bff7-476c-90f1-6c0d8112cc2e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258588 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4cll\" (UniqueName: \"kubernetes.io/projected/0aef8214-0bed-45cc-b3cd-77fcfbecb7aa-kube-api-access-p4cll\") pod \"catalog-operator-68c6474976-jr26c\" (UID: \"0aef8214-0bed-45cc-b3cd-77fcfbecb7aa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258611 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c8c049a4-28db-4ac4-807c-637c3837235a-node-bootstrap-token\") pod \"machine-config-server-l6nvl\" (UID: \"c8c049a4-28db-4ac4-807c-637c3837235a\") " pod="openshift-machine-config-operator/machine-config-server-l6nvl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258635 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c633e2b-4169-4277-a64b-85c249869b0c-webhook-cert\") pod \"packageserver-d55dfcdfc-f55bl\" (UID: \"0c633e2b-4169-4277-a64b-85c249869b0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258657 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a2888eb-7da9-4ef2-841b-22f170e66a54-serving-cert\") pod \"service-ca-operator-777779d784-phwq8\" (UID: \"9a2888eb-7da9-4ef2-841b-22f170e66a54\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-phwq8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258681 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgdgs\" (UniqueName: \"kubernetes.io/projected/9a2888eb-7da9-4ef2-841b-22f170e66a54-kube-api-access-qgdgs\") pod \"service-ca-operator-777779d784-phwq8\" (UID: \"9a2888eb-7da9-4ef2-841b-22f170e66a54\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-phwq8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258705 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8935f198-82ab-463c-96d5-300a23d613e6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s8mcw\" (UID: \"8935f198-82ab-463c-96d5-300a23d613e6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s8mcw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258738 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a3eedabc-4a51-4068-a9da-c8a867cb3219-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-6glnm\" (UID: \"a3eedabc-4a51-4068-a9da-c8a867cb3219\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6glnm" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258762 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0c633e2b-4169-4277-a64b-85c249869b0c-tmpfs\") pod \"packageserver-d55dfcdfc-f55bl\" (UID: \"0c633e2b-4169-4277-a64b-85c249869b0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258785 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b14853e5-2388-4dba-a3ca-dd9453ea3452-csi-data-dir\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258808 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7k47\" (UniqueName: \"kubernetes.io/projected/b14853e5-2388-4dba-a3ca-dd9453ea3452-kube-api-access-g7k47\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258885 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0ff5cea8-7b98-4a15-87cb-f07fb9cfabea-signing-cabundle\") pod \"service-ca-9c57cc56f-596cp\" (UID: \"0ff5cea8-7b98-4a15-87cb-f07fb9cfabea\") " pod="openshift-service-ca/service-ca-9c57cc56f-596cp" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258920 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b14853e5-2388-4dba-a3ca-dd9453ea3452-registration-dir\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258943 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6-proxy-tls\") pod \"machine-config-controller-84d6567774-bv9gv\" (UID: \"1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258966 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8935f198-82ab-463c-96d5-300a23d613e6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s8mcw\" (UID: \"8935f198-82ab-463c-96d5-300a23d613e6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s8mcw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.258999 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/faad9125-f494-49ae-ba77-941e6abe67b9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ltb69\" (UID: \"faad9125-f494-49ae-ba77-941e6abe67b9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ltb69" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259023 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c313309-b494-4209-a248-cb526aa88f36-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dh7z8\" (UID: \"2c313309-b494-4209-a248-cb526aa88f36\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dh7z8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259055 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9d0ddd49-ca6f-44aa-a3aa-5a7717845156-metrics-tls\") pod \"dns-operator-744455d44c-jvbsr\" (UID: \"9d0ddd49-ca6f-44aa-a3aa-5a7717845156\") " pod="openshift-dns-operator/dns-operator-744455d44c-jvbsr" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259074 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b14853e5-2388-4dba-a3ca-dd9453ea3452-socket-dir\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259094 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/53a413a7-34f8-4fb3-893b-c4fa9711017a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4dh8d\" (UID: \"53a413a7-34f8-4fb3-893b-c4fa9711017a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259113 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-config-volume\") pod \"collect-profiles-29412375-f596r\" (UID: \"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259137 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbvwx\" (UniqueName: \"kubernetes.io/projected/faad9125-f494-49ae-ba77-941e6abe67b9-kube-api-access-pbvwx\") pod \"control-plane-machine-set-operator-78cbb6b69f-ltb69\" (UID: \"faad9125-f494-49ae-ba77-941e6abe67b9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ltb69" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259158 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/503c8137-bff7-476c-90f1-6c0d8112cc2e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-dr7rf\" (UID: \"503c8137-bff7-476c-90f1-6c0d8112cc2e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259177 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8935f198-82ab-463c-96d5-300a23d613e6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s8mcw\" (UID: \"8935f198-82ab-463c-96d5-300a23d613e6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s8mcw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259197 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdv8l\" (UniqueName: \"kubernetes.io/projected/2bf3947e-3c18-4ef1-aca9-98419c558d17-kube-api-access-gdv8l\") pod \"ingress-operator-5b745b69d9-8jzsz\" (UID: \"2bf3947e-3c18-4ef1-aca9-98419c558d17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259249 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlm4m\" (UniqueName: \"kubernetes.io/projected/53a413a7-34f8-4fb3-893b-c4fa9711017a-kube-api-access-vlm4m\") pod \"marketplace-operator-79b997595-4dh8d\" (UID: \"53a413a7-34f8-4fb3-893b-c4fa9711017a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259289 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2bf3947e-3c18-4ef1-aca9-98419c558d17-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8jzsz\" (UID: \"2bf3947e-3c18-4ef1-aca9-98419c558d17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259328 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0ff5cea8-7b98-4a15-87cb-f07fb9cfabea-signing-key\") pod \"service-ca-9c57cc56f-596cp\" (UID: \"0ff5cea8-7b98-4a15-87cb-f07fb9cfabea\") " pod="openshift-service-ca/service-ca-9c57cc56f-596cp" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259357 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0aef8214-0bed-45cc-b3cd-77fcfbecb7aa-profile-collector-cert\") pod \"catalog-operator-68c6474976-jr26c\" (UID: \"0aef8214-0bed-45cc-b3cd-77fcfbecb7aa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259383 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c8c049a4-28db-4ac4-807c-637c3837235a-certs\") pod \"machine-config-server-l6nvl\" (UID: \"c8c049a4-28db-4ac4-807c-637c3837235a\") " pod="openshift-machine-config-operator/machine-config-server-l6nvl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259407 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9nsd\" (UniqueName: \"kubernetes.io/projected/c8c049a4-28db-4ac4-807c-637c3837235a-kube-api-access-n9nsd\") pod \"machine-config-server-l6nvl\" (UID: \"c8c049a4-28db-4ac4-807c-637c3837235a\") " pod="openshift-machine-config-operator/machine-config-server-l6nvl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259475 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwrzj\" (UniqueName: \"kubernetes.io/projected/0ff5cea8-7b98-4a15-87cb-f07fb9cfabea-kube-api-access-nwrzj\") pod \"service-ca-9c57cc56f-596cp\" (UID: \"0ff5cea8-7b98-4a15-87cb-f07fb9cfabea\") " pod="openshift-service-ca/service-ca-9c57cc56f-596cp" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259529 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q4s7\" (UniqueName: \"kubernetes.io/projected/8b9018a9-c76c-4471-8b27-06c1cbc065eb-kube-api-access-4q4s7\") pod \"multus-admission-controller-857f4d67dd-s8vm2\" (UID: \"8b9018a9-c76c-4471-8b27-06c1cbc065eb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s8vm2" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259551 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a2888eb-7da9-4ef2-841b-22f170e66a54-config\") pod \"service-ca-operator-777779d784-phwq8\" (UID: \"9a2888eb-7da9-4ef2-841b-22f170e66a54\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-phwq8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259578 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c633e2b-4169-4277-a64b-85c249869b0c-apiservice-cert\") pod \"packageserver-d55dfcdfc-f55bl\" (UID: \"0c633e2b-4169-4277-a64b-85c249869b0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259598 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5cdd\" (UniqueName: \"kubernetes.io/projected/0c633e2b-4169-4277-a64b-85c249869b0c-kube-api-access-r5cdd\") pod \"packageserver-d55dfcdfc-f55bl\" (UID: \"0c633e2b-4169-4277-a64b-85c249869b0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259634 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-secret-volume\") pod \"collect-profiles-29412375-f596r\" (UID: \"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" Dec 03 06:29:59 crc kubenswrapper[4818]: E1203 06:29:59.259695 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:29:59.759648186 +0000 UTC m=+157.451256948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259724 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2bf3947e-3c18-4ef1-aca9-98419c558d17-trusted-ca\") pod \"ingress-operator-5b745b69d9-8jzsz\" (UID: \"2bf3947e-3c18-4ef1-aca9-98419c558d17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259776 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c313309-b494-4209-a248-cb526aa88f36-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dh7z8\" (UID: \"2c313309-b494-4209-a248-cb526aa88f36\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dh7z8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259808 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxjhq\" (UniqueName: \"kubernetes.io/projected/1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6-kube-api-access-xxjhq\") pod \"machine-config-controller-84d6567774-bv9gv\" (UID: \"1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259877 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53a413a7-34f8-4fb3-893b-c4fa9711017a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4dh8d\" (UID: \"53a413a7-34f8-4fb3-893b-c4fa9711017a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259902 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flzcx\" (UniqueName: \"kubernetes.io/projected/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-kube-api-access-flzcx\") pod \"collect-profiles-29412375-f596r\" (UID: \"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.259963 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/503c8137-bff7-476c-90f1-6c0d8112cc2e-srv-cert\") pod \"olm-operator-6b444d44fb-dr7rf\" (UID: \"503c8137-bff7-476c-90f1-6c0d8112cc2e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.260012 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xngvs\" (UniqueName: \"kubernetes.io/projected/a0dc41b5-d81d-4a47-aaad-8a9d8216e513-kube-api-access-xngvs\") pod \"ingress-canary-nmqf4\" (UID: \"a0dc41b5-d81d-4a47-aaad-8a9d8216e513\") " pod="openshift-ingress-canary/ingress-canary-nmqf4" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.260043 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwh6f\" (UniqueName: \"kubernetes.io/projected/2c313309-b494-4209-a248-cb526aa88f36-kube-api-access-qwh6f\") pod \"kube-storage-version-migrator-operator-b67b599dd-dh7z8\" (UID: \"2c313309-b494-4209-a248-cb526aa88f36\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dh7z8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.261340 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b14853e5-2388-4dba-a3ca-dd9453ea3452-mountpoint-dir\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.261859 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2bf3947e-3c18-4ef1-aca9-98419c558d17-trusted-ca\") pod \"ingress-operator-5b745b69d9-8jzsz\" (UID: \"2bf3947e-3c18-4ef1-aca9-98419c558d17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.262510 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53a413a7-34f8-4fb3-893b-c4fa9711017a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4dh8d\" (UID: \"53a413a7-34f8-4fb3-893b-c4fa9711017a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.263284 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a63ab77-ab0e-4f43-9d21-2ec684236ce9-config-volume\") pod \"dns-default-79bwt\" (UID: \"3a63ab77-ab0e-4f43-9d21-2ec684236ce9\") " pod="openshift-dns/dns-default-79bwt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.263430 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77ws5\" (UniqueName: \"kubernetes.io/projected/a45cf1f5-9679-484d-920c-01e9a0d24587-kube-api-access-77ws5\") pod \"console-operator-58897d9998-2sn7n\" (UID: \"a45cf1f5-9679-484d-920c-01e9a0d24587\") " pod="openshift-console-operator/console-operator-58897d9998-2sn7n" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.263705 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c313309-b494-4209-a248-cb526aa88f36-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dh7z8\" (UID: \"2c313309-b494-4209-a248-cb526aa88f36\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dh7z8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.263824 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-bv9gv\" (UID: \"1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.264718 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b14853e5-2388-4dba-a3ca-dd9453ea3452-csi-data-dir\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.264765 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a2888eb-7da9-4ef2-841b-22f170e66a54-config\") pod \"service-ca-operator-777779d784-phwq8\" (UID: \"9a2888eb-7da9-4ef2-841b-22f170e66a54\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-phwq8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.264800 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b14853e5-2388-4dba-a3ca-dd9453ea3452-registration-dir\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.265633 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0ff5cea8-7b98-4a15-87cb-f07fb9cfabea-signing-cabundle\") pod \"service-ca-9c57cc56f-596cp\" (UID: \"0ff5cea8-7b98-4a15-87cb-f07fb9cfabea\") " pod="openshift-service-ca/service-ca-9c57cc56f-596cp" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.265698 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a3eedabc-4a51-4068-a9da-c8a867cb3219-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-6glnm\" (UID: \"a3eedabc-4a51-4068-a9da-c8a867cb3219\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6glnm" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.266091 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8935f198-82ab-463c-96d5-300a23d613e6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s8mcw\" (UID: \"8935f198-82ab-463c-96d5-300a23d613e6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s8mcw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.267280 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b14853e5-2388-4dba-a3ca-dd9453ea3452-socket-dir\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.267455 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0c633e2b-4169-4277-a64b-85c249869b0c-tmpfs\") pod \"packageserver-d55dfcdfc-f55bl\" (UID: \"0c633e2b-4169-4277-a64b-85c249869b0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.269562 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0aef8214-0bed-45cc-b3cd-77fcfbecb7aa-profile-collector-cert\") pod \"catalog-operator-68c6474976-jr26c\" (UID: \"0aef8214-0bed-45cc-b3cd-77fcfbecb7aa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.269730 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/503c8137-bff7-476c-90f1-6c0d8112cc2e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-dr7rf\" (UID: \"503c8137-bff7-476c-90f1-6c0d8112cc2e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.270056 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3a63ab77-ab0e-4f43-9d21-2ec684236ce9-metrics-tls\") pod \"dns-default-79bwt\" (UID: \"3a63ab77-ab0e-4f43-9d21-2ec684236ce9\") " pod="openshift-dns/dns-default-79bwt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.270179 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2bf3947e-3c18-4ef1-aca9-98419c558d17-metrics-tls\") pod \"ingress-operator-5b745b69d9-8jzsz\" (UID: \"2bf3947e-3c18-4ef1-aca9-98419c558d17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.270667 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c633e2b-4169-4277-a64b-85c249869b0c-apiservice-cert\") pod \"packageserver-d55dfcdfc-f55bl\" (UID: \"0c633e2b-4169-4277-a64b-85c249869b0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.270675 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8b9018a9-c76c-4471-8b27-06c1cbc065eb-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-s8vm2\" (UID: \"8b9018a9-c76c-4471-8b27-06c1cbc065eb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s8vm2" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.270994 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c633e2b-4169-4277-a64b-85c249869b0c-webhook-cert\") pod \"packageserver-d55dfcdfc-f55bl\" (UID: \"0c633e2b-4169-4277-a64b-85c249869b0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.271737 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0ff5cea8-7b98-4a15-87cb-f07fb9cfabea-signing-key\") pod \"service-ca-9c57cc56f-596cp\" (UID: \"0ff5cea8-7b98-4a15-87cb-f07fb9cfabea\") " pod="openshift-service-ca/service-ca-9c57cc56f-596cp" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.271991 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6-proxy-tls\") pod \"machine-config-controller-84d6567774-bv9gv\" (UID: \"1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.272263 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0aef8214-0bed-45cc-b3cd-77fcfbecb7aa-srv-cert\") pod \"catalog-operator-68c6474976-jr26c\" (UID: \"0aef8214-0bed-45cc-b3cd-77fcfbecb7aa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.272535 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-secret-volume\") pod \"collect-profiles-29412375-f596r\" (UID: \"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.272903 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/faad9125-f494-49ae-ba77-941e6abe67b9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ltb69\" (UID: \"faad9125-f494-49ae-ba77-941e6abe67b9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ltb69" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.272912 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0dc41b5-d81d-4a47-aaad-8a9d8216e513-cert\") pod \"ingress-canary-nmqf4\" (UID: \"a0dc41b5-d81d-4a47-aaad-8a9d8216e513\") " pod="openshift-ingress-canary/ingress-canary-nmqf4" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.273163 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-config-volume\") pod \"collect-profiles-29412375-f596r\" (UID: \"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.274317 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a2888eb-7da9-4ef2-841b-22f170e66a54-serving-cert\") pod \"service-ca-operator-777779d784-phwq8\" (UID: \"9a2888eb-7da9-4ef2-841b-22f170e66a54\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-phwq8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.274468 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/503c8137-bff7-476c-90f1-6c0d8112cc2e-srv-cert\") pod \"olm-operator-6b444d44fb-dr7rf\" (UID: \"503c8137-bff7-476c-90f1-6c0d8112cc2e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.274777 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9d0ddd49-ca6f-44aa-a3aa-5a7717845156-metrics-tls\") pod \"dns-operator-744455d44c-jvbsr\" (UID: \"9d0ddd49-ca6f-44aa-a3aa-5a7717845156\") " pod="openshift-dns-operator/dns-operator-744455d44c-jvbsr" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.275054 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/53a413a7-34f8-4fb3-893b-c4fa9711017a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4dh8d\" (UID: \"53a413a7-34f8-4fb3-893b-c4fa9711017a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.276212 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8935f198-82ab-463c-96d5-300a23d613e6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s8mcw\" (UID: \"8935f198-82ab-463c-96d5-300a23d613e6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s8mcw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.277763 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c313309-b494-4209-a248-cb526aa88f36-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dh7z8\" (UID: \"2c313309-b494-4209-a248-cb526aa88f36\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dh7z8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.278080 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c8c049a4-28db-4ac4-807c-637c3837235a-certs\") pod \"machine-config-server-l6nvl\" (UID: \"c8c049a4-28db-4ac4-807c-637c3837235a\") " pod="openshift-machine-config-operator/machine-config-server-l6nvl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.279080 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c8c049a4-28db-4ac4-807c-637c3837235a-node-bootstrap-token\") pod \"machine-config-server-l6nvl\" (UID: \"c8c049a4-28db-4ac4-807c-637c3837235a\") " pod="openshift-machine-config-operator/machine-config-server-l6nvl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.284494 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxwvp\" (UniqueName: \"kubernetes.io/projected/10f43944-25cf-4ae7-8298-82635d81b62e-kube-api-access-nxwvp\") pod \"openshift-controller-manager-operator-756b6f6bc6-l6mbg\" (UID: \"10f43944-25cf-4ae7-8298-82635d81b62e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6mbg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.296461 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whxj7\" (UniqueName: \"kubernetes.io/projected/a688a250-d68c-4995-b6ff-85daf52874ac-kube-api-access-whxj7\") pod \"authentication-operator-69f744f599-rl64s\" (UID: \"a688a250-d68c-4995-b6ff-85daf52874ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.302848 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tx2n9"] Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.318939 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k4pkt\" (UID: \"aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt" Dec 03 06:29:59 crc kubenswrapper[4818]: W1203 06:29:59.333399 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85223238_8e03_42bb_8790_8887e80d0e52.slice/crio-5fd0e563760a66428d2b413c20f88023c94ef9599786e7b35c68ceb061f80b5d WatchSource:0}: Error finding container 5fd0e563760a66428d2b413c20f88023c94ef9599786e7b35c68ceb061f80b5d: Status 404 returned error can't find the container with id 5fd0e563760a66428d2b413c20f88023c94ef9599786e7b35c68ceb061f80b5d Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.336365 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp6x8\" (UniqueName: \"kubernetes.io/projected/a9ca675e-e374-467c-aa38-da781b84e3dd-kube-api-access-rp6x8\") pod \"machine-config-operator-74547568cd-cnpz5\" (UID: \"a9ca675e-e374-467c-aa38-da781b84e3dd\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.360757 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: E1203 06:29:59.361184 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:29:59.861166748 +0000 UTC m=+157.552775560 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.364102 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86x4d\" (UniqueName: \"kubernetes.io/projected/dae2df0a-0703-40b5-9040-04b3af885585-kube-api-access-86x4d\") pod \"cluster-samples-operator-665b6dd947-j8d4t\" (UID: \"dae2df0a-0703-40b5-9040-04b3af885585\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j8d4t" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.375028 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.378259 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.386640 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/06af2589-0701-4b16-a31c-eb39fe1c8b99-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-j25z9\" (UID: \"06af2589-0701-4b16-a31c-eb39fe1c8b99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j25z9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.399282 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px568\" (UniqueName: \"kubernetes.io/projected/39dce7bf-1502-40fa-9d9c-41378ccad702-kube-api-access-px568\") pod \"apiserver-76f77b778f-qnxnn\" (UID: \"39dce7bf-1502-40fa-9d9c-41378ccad702\") " pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.407235 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-2sn7n" Dec 03 06:29:59 crc kubenswrapper[4818]: W1203 06:29:59.411528 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90705934_7171_4811_85d8_8769a47f7fc7.slice/crio-7985e797308d9630e1c7f70b3b0b35fd2f22b4a209e4f35914193a2793529bc6 WatchSource:0}: Error finding container 7985e797308d9630e1c7f70b3b0b35fd2f22b4a209e4f35914193a2793529bc6: Status 404 returned error can't find the container with id 7985e797308d9630e1c7f70b3b0b35fd2f22b4a209e4f35914193a2793529bc6 Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.418255 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gt4k9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.418773 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnmvg\" (UniqueName: \"kubernetes.io/projected/3dbb4222-9650-4663-95bf-bd2fa4f4e678-kube-api-access-cnmvg\") pod \"openshift-config-operator-7777fb866f-h6fkg\" (UID: \"3dbb4222-9650-4663-95bf-bd2fa4f4e678\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.427926 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fv5lc"] Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.438760 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnlc6\" (UniqueName: \"kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-kube-api-access-gnlc6\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.459349 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rp5w\" (UniqueName: \"kubernetes.io/projected/05f7659c-85e9-4337-8741-9958f12895d4-kube-api-access-6rp5w\") pod \"router-default-5444994796-fz577\" (UID: \"05f7659c-85e9-4337-8741-9958f12895d4\") " pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.461554 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg" Dec 03 06:29:59 crc kubenswrapper[4818]: E1203 06:29:59.461625 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:29:59.961605613 +0000 UTC m=+157.653214365 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.461561 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.461948 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: E1203 06:29:59.462434 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:29:59.962424254 +0000 UTC m=+157.654033006 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.465569 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6mbg" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.482648 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkks8\" (UniqueName: \"kubernetes.io/projected/6d2839d4-e624-4d61-9225-894a998e9e9e-kube-api-access-vkks8\") pod \"console-f9d7485db-sccrw\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.497298 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" event={"ID":"d7f12c1b-f504-4155-bd29-bcb6ab57b646","Type":"ContainerStarted","Data":"4bf926d053399f203eb58cfbe643d78e35f388e3f6483f5cce576a5358ebb2ee"} Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.499172 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tx2n9" event={"ID":"85223238-8e03-42bb-8790-8887e80d0e52","Type":"ContainerStarted","Data":"1268ba119aa037b5c38d5fd36fff58ba450a54d84bec55a6faacb828a04fab46"} Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.499196 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tx2n9" event={"ID":"85223238-8e03-42bb-8790-8887e80d0e52","Type":"ContainerStarted","Data":"5fd0e563760a66428d2b413c20f88023c94ef9599786e7b35c68ceb061f80b5d"} Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.499998 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j25z9" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.501624 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msqs7\" (UniqueName: \"kubernetes.io/projected/4b90f9f4-fcfa-46b3-80de-b8e335c84e5b-kube-api-access-msqs7\") pod \"downloads-7954f5f757-pvb2q\" (UID: \"4b90f9f4-fcfa-46b3-80de-b8e335c84e5b\") " pod="openshift-console/downloads-7954f5f757-pvb2q" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.503888 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" event={"ID":"6a406cec-8381-4a82-8d46-3ababcda72b4","Type":"ContainerStarted","Data":"8941867e59b8f0abdf51e22f253e435eea713ca60abc759608fac0bbc364f02a"} Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.503920 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" event={"ID":"6a406cec-8381-4a82-8d46-3ababcda72b4","Type":"ContainerStarted","Data":"e8006f24ea9ac229abfd1f35c197ad198a1e9e8c896924f87daba48046e1d43d"} Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.504708 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.517672 4818 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-5jw58 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.517709 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" podUID="6a406cec-8381-4a82-8d46-3ababcda72b4" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.534508 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vr5nb\" (UniqueName: \"kubernetes.io/projected/d167fcbe-c833-43f2-ada0-885857d568a9-kube-api-access-vr5nb\") pod \"etcd-operator-b45778765-6h7lb\" (UID: \"d167fcbe-c833-43f2-ada0-885857d568a9\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.541413 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" event={"ID":"8ec6e899-8380-4812-8f65-6fd72db12939","Type":"ContainerStarted","Data":"dc26bc1346b1308743d97991bbace0d73ffaa6b85ae44cc07a7c111709ef377f"} Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.541454 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" event={"ID":"8ec6e899-8380-4812-8f65-6fd72db12939","Type":"ContainerStarted","Data":"80b0ad4372e370559120bd0151c3497c288cf88d9110c4971a0594b7dff794aa"} Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.541463 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" event={"ID":"8ec6e899-8380-4812-8f65-6fd72db12939","Type":"ContainerStarted","Data":"fa1e89355a1b8c4178c8645721792a2db392c13d72bc700b9c7ea1b533ab1fbd"} Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.544806 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9s47\" (UniqueName: \"kubernetes.io/projected/aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721-kube-api-access-b9s47\") pod \"cluster-image-registry-operator-dc59b4c8b-k4pkt\" (UID: \"aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.551327 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.555454 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" event={"ID":"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7","Type":"ContainerStarted","Data":"37f806687d01927c0c9fe9c57dd2e12b9ce2f0a32b8287d06ea0ca1993b5d98b"} Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.555490 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" event={"ID":"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7","Type":"ContainerStarted","Data":"86a0a80e10db78e3141bdd3d5232448b580b394eedcf1c04b18d7d80ebc41fc5"} Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.556365 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.564090 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:29:59 crc kubenswrapper[4818]: E1203 06:29:59.565202 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:00.065185717 +0000 UTC m=+157.756794469 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.579235 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" event={"ID":"90705934-7171-4811-85d8-8769a47f7fc7","Type":"ContainerStarted","Data":"7985e797308d9630e1c7f70b3b0b35fd2f22b4a209e4f35914193a2793529bc6"} Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.585588 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp4sk\" (UniqueName: \"kubernetes.io/projected/9d0ddd49-ca6f-44aa-a3aa-5a7717845156-kube-api-access-cp4sk\") pod \"dns-operator-744455d44c-jvbsr\" (UID: \"9d0ddd49-ca6f-44aa-a3aa-5a7717845156\") " pod="openshift-dns-operator/dns-operator-744455d44c-jvbsr" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.589050 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" event={"ID":"0122f472-9da5-40c1-aa0a-ab6219543b1d","Type":"ContainerStarted","Data":"af8727ba582d813ac43ac36611b3c14549e3f2e94d3ee1ef960694a523dc551c"} Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.590344 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jvbsr" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.596291 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b9c2\" (UniqueName: \"kubernetes.io/projected/3deb8606-51f3-4a91-bcfd-0533f315e829-kube-api-access-8b9c2\") pod \"migrator-59844c95c7-8x7mv\" (UID: \"3deb8606-51f3-4a91-bcfd-0533f315e829\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8x7mv" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.610514 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.627766 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rl64s"] Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.630962 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwh6f\" (UniqueName: \"kubernetes.io/projected/2c313309-b494-4209-a248-cb526aa88f36-kube-api-access-qwh6f\") pod \"kube-storage-version-migrator-operator-b67b599dd-dh7z8\" (UID: \"2c313309-b494-4209-a248-cb526aa88f36\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dh7z8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.634352 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.650847 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flzcx\" (UniqueName: \"kubernetes.io/projected/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-kube-api-access-flzcx\") pod \"collect-profiles-29412375-f596r\" (UID: \"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.655803 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:29:59 crc kubenswrapper[4818]: W1203 06:29:59.662460 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda688a250_d68c_4995_b6ff_85daf52874ac.slice/crio-af26d1ca60830b4609e603f5ddf0feb56bfcf4ad71a0c289489d57f889491cfe WatchSource:0}: Error finding container af26d1ca60830b4609e603f5ddf0feb56bfcf4ad71a0c289489d57f889491cfe: Status 404 returned error can't find the container with id af26d1ca60830b4609e603f5ddf0feb56bfcf4ad71a0c289489d57f889491cfe Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.675175 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j8d4t" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.675775 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: E1203 06:29:59.678711 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:00.178697216 +0000 UTC m=+157.870305968 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.694115 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xngvs\" (UniqueName: \"kubernetes.io/projected/a0dc41b5-d81d-4a47-aaad-8a9d8216e513-kube-api-access-xngvs\") pod \"ingress-canary-nmqf4\" (UID: \"a0dc41b5-d81d-4a47-aaad-8a9d8216e513\") " pod="openshift-ingress-canary/ingress-canary-nmqf4" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.694391 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.695009 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2sn7n"] Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.701112 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.709280 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbvwx\" (UniqueName: \"kubernetes.io/projected/faad9125-f494-49ae-ba77-941e6abe67b9-kube-api-access-pbvwx\") pod \"control-plane-machine-set-operator-78cbb6b69f-ltb69\" (UID: \"faad9125-f494-49ae-ba77-941e6abe67b9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ltb69" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.729390 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxjhq\" (UniqueName: \"kubernetes.io/projected/1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6-kube-api-access-xxjhq\") pod \"machine-config-controller-84d6567774-bv9gv\" (UID: \"1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.732126 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq877\" (UniqueName: \"kubernetes.io/projected/a3eedabc-4a51-4068-a9da-c8a867cb3219-kube-api-access-vq877\") pod \"package-server-manager-789f6589d5-6glnm\" (UID: \"a3eedabc-4a51-4068-a9da-c8a867cb3219\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6glnm" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.765352 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlm4m\" (UniqueName: \"kubernetes.io/projected/53a413a7-34f8-4fb3-893b-c4fa9711017a-kube-api-access-vlm4m\") pod \"marketplace-operator-79b997595-4dh8d\" (UID: \"53a413a7-34f8-4fb3-893b-c4fa9711017a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.767774 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdv8l\" (UniqueName: \"kubernetes.io/projected/2bf3947e-3c18-4ef1-aca9-98419c558d17-kube-api-access-gdv8l\") pod \"ingress-operator-5b745b69d9-8jzsz\" (UID: \"2bf3947e-3c18-4ef1-aca9-98419c558d17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.775626 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-pvb2q" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.777390 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:29:59 crc kubenswrapper[4818]: E1203 06:29:59.778941 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:00.278918736 +0000 UTC m=+157.970527488 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.780124 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2bf3947e-3c18-4ef1-aca9-98419c558d17-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8jzsz\" (UID: \"2bf3947e-3c18-4ef1-aca9-98419c558d17\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.798460 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q4s7\" (UniqueName: \"kubernetes.io/projected/8b9018a9-c76c-4471-8b27-06c1cbc065eb-kube-api-access-4q4s7\") pod \"multus-admission-controller-857f4d67dd-s8vm2\" (UID: \"8b9018a9-c76c-4471-8b27-06c1cbc065eb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s8vm2" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.821173 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.823778 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gt4k9"] Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.835090 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9nsd\" (UniqueName: \"kubernetes.io/projected/c8c049a4-28db-4ac4-807c-637c3837235a-kube-api-access-n9nsd\") pod \"machine-config-server-l6nvl\" (UID: \"c8c049a4-28db-4ac4-807c-637c3837235a\") " pod="openshift-machine-config-operator/machine-config-server-l6nvl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.866123 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dh7z8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.867711 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwrzj\" (UniqueName: \"kubernetes.io/projected/0ff5cea8-7b98-4a15-87cb-f07fb9cfabea-kube-api-access-nwrzj\") pod \"service-ca-9c57cc56f-596cp\" (UID: \"0ff5cea8-7b98-4a15-87cb-f07fb9cfabea\") " pod="openshift-service-ca/service-ca-9c57cc56f-596cp" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.872375 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7k47\" (UniqueName: \"kubernetes.io/projected/b14853e5-2388-4dba-a3ca-dd9453ea3452-kube-api-access-g7k47\") pod \"csi-hostpathplugin-gkgqw\" (UID: \"b14853e5-2388-4dba-a3ca-dd9453ea3452\") " pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.875922 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8x7mv" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.877324 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5cdd\" (UniqueName: \"kubernetes.io/projected/0c633e2b-4169-4277-a64b-85c249869b0c-kube-api-access-r5cdd\") pod \"packageserver-d55dfcdfc-f55bl\" (UID: \"0c633e2b-4169-4277-a64b-85c249869b0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.887697 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.887971 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:29:59 crc kubenswrapper[4818]: E1203 06:29:59.889084 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:00.389052341 +0000 UTC m=+158.080661093 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.896566 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.908339 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ltb69" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.912570 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.919619 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-596cp" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.940691 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtgcj\" (UniqueName: \"kubernetes.io/projected/3a63ab77-ab0e-4f43-9d21-2ec684236ce9-kube-api-access-dtgcj\") pod \"dns-default-79bwt\" (UID: \"3a63ab77-ab0e-4f43-9d21-2ec684236ce9\") " pod="openshift-dns/dns-default-79bwt" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.943735 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4cll\" (UniqueName: \"kubernetes.io/projected/0aef8214-0bed-45cc-b3cd-77fcfbecb7aa-kube-api-access-p4cll\") pod \"catalog-operator-68c6474976-jr26c\" (UID: \"0aef8214-0bed-45cc-b3cd-77fcfbecb7aa\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.945983 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.949468 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fcz6\" (UniqueName: \"kubernetes.io/projected/503c8137-bff7-476c-90f1-6c0d8112cc2e-kube-api-access-2fcz6\") pod \"olm-operator-6b444d44fb-dr7rf\" (UID: \"503c8137-bff7-476c-90f1-6c0d8112cc2e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.954256 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-s8vm2" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.963294 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6glnm" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.980462 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-l6nvl" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.983143 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8935f198-82ab-463c-96d5-300a23d613e6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s8mcw\" (UID: \"8935f198-82ab-463c-96d5-300a23d613e6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s8mcw" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.987223 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgdgs\" (UniqueName: \"kubernetes.io/projected/9a2888eb-7da9-4ef2-841b-22f170e66a54-kube-api-access-qgdgs\") pod \"service-ca-operator-777779d784-phwq8\" (UID: \"9a2888eb-7da9-4ef2-841b-22f170e66a54\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-phwq8" Dec 03 06:29:59 crc kubenswrapper[4818]: I1203 06:29:59.987612 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-nmqf4" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.002723 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:00 crc kubenswrapper[4818]: E1203 06:30:00.003410 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:00.503393972 +0000 UTC m=+158.195002724 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.003672 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-79bwt" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.014192 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.093778 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg"] Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.105483 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:00 crc kubenswrapper[4818]: E1203 06:30:00.105907 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:00.605895598 +0000 UTC m=+158.297504350 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.132380 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r"] Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.145173 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s8mcw" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.158367 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.176529 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9"] Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.177610 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9"] Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.177698 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.206958 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:00 crc kubenswrapper[4818]: E1203 06:30:00.207344 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:00.707329478 +0000 UTC m=+158.398938230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.229056 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.236396 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.271982 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-phwq8" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.309048 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f033be5e-5c71-4aeb-8825-bcf234d174a1-secret-volume\") pod \"collect-profiles-29412390-c45b9\" (UID: \"f033be5e-5c71-4aeb-8825-bcf234d174a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.309231 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:00 crc kubenswrapper[4818]: E1203 06:30:00.309565 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:00.809551447 +0000 UTC m=+158.501160209 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.310027 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f033be5e-5c71-4aeb-8825-bcf234d174a1-config-volume\") pod \"collect-profiles-29412390-c45b9\" (UID: \"f033be5e-5c71-4aeb-8825-bcf234d174a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.310143 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfmch\" (UniqueName: \"kubernetes.io/projected/f033be5e-5c71-4aeb-8825-bcf234d174a1-kube-api-access-lfmch\") pod \"collect-profiles-29412390-c45b9\" (UID: \"f033be5e-5c71-4aeb-8825-bcf234d174a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.412844 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:00 crc kubenswrapper[4818]: E1203 06:30:00.412979 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:00.912949366 +0000 UTC m=+158.604558118 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.413670 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f033be5e-5c71-4aeb-8825-bcf234d174a1-secret-volume\") pod \"collect-profiles-29412390-c45b9\" (UID: \"f033be5e-5c71-4aeb-8825-bcf234d174a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.413750 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.413790 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f033be5e-5c71-4aeb-8825-bcf234d174a1-config-volume\") pod \"collect-profiles-29412390-c45b9\" (UID: \"f033be5e-5c71-4aeb-8825-bcf234d174a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.413851 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfmch\" (UniqueName: \"kubernetes.io/projected/f033be5e-5c71-4aeb-8825-bcf234d174a1-kube-api-access-lfmch\") pod \"collect-profiles-29412390-c45b9\" (UID: \"f033be5e-5c71-4aeb-8825-bcf234d174a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" Dec 03 06:30:00 crc kubenswrapper[4818]: E1203 06:30:00.414551 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:00.914539355 +0000 UTC m=+158.606148207 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.416025 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f033be5e-5c71-4aeb-8825-bcf234d174a1-config-volume\") pod \"collect-profiles-29412390-c45b9\" (UID: \"f033be5e-5c71-4aeb-8825-bcf234d174a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.445171 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f033be5e-5c71-4aeb-8825-bcf234d174a1-secret-volume\") pod \"collect-profiles-29412390-c45b9\" (UID: \"f033be5e-5c71-4aeb-8825-bcf234d174a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.488737 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfmch\" (UniqueName: \"kubernetes.io/projected/f033be5e-5c71-4aeb-8825-bcf234d174a1-kube-api-access-lfmch\") pod \"collect-profiles-29412390-c45b9\" (UID: \"f033be5e-5c71-4aeb-8825-bcf234d174a1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.514624 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:00 crc kubenswrapper[4818]: E1203 06:30:00.516665 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:01.016647412 +0000 UTC m=+158.708256164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.617530 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:00 crc kubenswrapper[4818]: E1203 06:30:00.619036 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:01.119019585 +0000 UTC m=+158.810628337 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.619041 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" event={"ID":"90705934-7171-4811-85d8-8769a47f7fc7","Type":"ContainerStarted","Data":"c71c399f5ba736182bb961347b734daf83f1c2c6ce135d2ffd13a167cf489586"} Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.639990 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" event={"ID":"a688a250-d68c-4995-b6ff-85daf52874ac","Type":"ContainerStarted","Data":"1ff06a1bf965989a5e71b7735e263f2d7eba85a0e1dd19189892485d6c192925"} Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.640033 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" event={"ID":"a688a250-d68c-4995-b6ff-85daf52874ac","Type":"ContainerStarted","Data":"af26d1ca60830b4609e603f5ddf0feb56bfcf4ad71a0c289489d57f889491cfe"} Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.644988 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-2sn7n" event={"ID":"a45cf1f5-9679-484d-920c-01e9a0d24587","Type":"ContainerStarted","Data":"891a7dd7d1aa90915fb411f969862f34dc80ea5242d133908e333261eda48c55"} Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.645027 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-2sn7n" event={"ID":"a45cf1f5-9679-484d-920c-01e9a0d24587","Type":"ContainerStarted","Data":"29846603219ef999f35483081bbdda2f045e14b9ff491696ed4fa1fb3373da64"} Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.645459 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-2sn7n" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.653394 4818 generic.go:334] "Generic (PLEG): container finished" podID="0122f472-9da5-40c1-aa0a-ab6219543b1d" containerID="d47be17af16b7dba36a389c4075f33bbf8cdd33f2fee2858e50ef7284695462a" exitCode=0 Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.653488 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" event={"ID":"0122f472-9da5-40c1-aa0a-ab6219543b1d","Type":"ContainerDied","Data":"d47be17af16b7dba36a389c4075f33bbf8cdd33f2fee2858e50ef7284695462a"} Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.660007 4818 patch_prober.go:28] interesting pod/console-operator-58897d9998-2sn7n container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.660060 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-2sn7n" podUID="a45cf1f5-9679-484d-920c-01e9a0d24587" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.664725 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tx2n9" podStartSLOduration=135.66470561 podStartE2EDuration="2m15.66470561s" podCreationTimestamp="2025-12-03 06:27:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:00.625326042 +0000 UTC m=+158.316934794" watchObservedRunningTime="2025-12-03 06:30:00.66470561 +0000 UTC m=+158.356314362" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.692444 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-l6nvl" event={"ID":"c8c049a4-28db-4ac4-807c-637c3837235a","Type":"ContainerStarted","Data":"1cdd79a9ba3bdc3c50488d1025c1bc6316d76960035c134c656163772995bfe1"} Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.692529 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.712622 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-fz577" event={"ID":"05f7659c-85e9-4337-8741-9958f12895d4","Type":"ContainerStarted","Data":"f615ed53e57c7b24a652d2dfba69df8b5378b96240cff1141aa1dbaa258543e5"} Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.712661 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-fz577" event={"ID":"05f7659c-85e9-4337-8741-9958f12895d4","Type":"ContainerStarted","Data":"a9189bbd9db07fc64b342b0be0551c45dcdc28d06346f4cc22adc0a2aa3aeb7c"} Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.714258 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gt4k9" event={"ID":"b6220eaa-3a5d-4c00-8091-fad092c5c77d","Type":"ContainerStarted","Data":"5da4e5baa8ff6168b78fcbcc23f785f6edaf55ebdf2f3f8bee8adb7b8e3aa43e"} Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.717866 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg" event={"ID":"3dbb4222-9650-4663-95bf-bd2fa4f4e678","Type":"ContainerStarted","Data":"62dbe43df0df5d3446a9c5f3b9172eedec57d822f75f287cb2a6f627930c887f"} Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.719389 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:00 crc kubenswrapper[4818]: E1203 06:30:00.719624 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:01.219601474 +0000 UTC m=+158.911210226 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.719698 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:00 crc kubenswrapper[4818]: E1203 06:30:00.720024 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:01.220012834 +0000 UTC m=+158.911621586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.761647 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" event={"ID":"d7f12c1b-f504-4155-bd29-bcb6ab57b646","Type":"ContainerStarted","Data":"8cca0fe7a690cbf136550efe0435f181a102b59974fe93863f2476a130111250"} Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.761693 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.789604 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.820988 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:00 crc kubenswrapper[4818]: E1203 06:30:00.824675 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:01.324653323 +0000 UTC m=+159.016262125 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.844187 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5"] Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.925096 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:00 crc kubenswrapper[4818]: E1203 06:30:00.926335 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:01.426319839 +0000 UTC m=+159.117928591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.945488 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-jf227" podStartSLOduration=134.945466615 podStartE2EDuration="2m14.945466615s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:00.933091697 +0000 UTC m=+158.624700459" watchObservedRunningTime="2025-12-03 06:30:00.945466615 +0000 UTC m=+158.637075367" Dec 03 06:30:00 crc kubenswrapper[4818]: I1203 06:30:00.986386 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j25z9"] Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.026978 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:01 crc kubenswrapper[4818]: E1203 06:30:01.027359 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:01.527342649 +0000 UTC m=+159.218951401 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.044073 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6mbg"] Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.128800 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:01 crc kubenswrapper[4818]: E1203 06:30:01.129448 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:01.629431965 +0000 UTC m=+159.321040717 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.230344 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:01 crc kubenswrapper[4818]: E1203 06:30:01.230695 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:01.73068004 +0000 UTC m=+159.422288792 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.331752 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:01 crc kubenswrapper[4818]: E1203 06:30:01.332189 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:01.832175831 +0000 UTC m=+159.523784583 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.414273 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" podStartSLOduration=135.41425378 podStartE2EDuration="2m15.41425378s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:01.39168596 +0000 UTC m=+159.083294712" watchObservedRunningTime="2025-12-03 06:30:01.41425378 +0000 UTC m=+159.105862532" Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.438522 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:01 crc kubenswrapper[4818]: E1203 06:30:01.438615 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:01.938593445 +0000 UTC m=+159.630202197 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.438773 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:01 crc kubenswrapper[4818]: E1203 06:30:01.439093 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:01.939086747 +0000 UTC m=+159.630695499 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.461370 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.539392 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:01 crc kubenswrapper[4818]: E1203 06:30:01.540116 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:02.040098026 +0000 UTC m=+159.731706778 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.552171 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" podStartSLOduration=135.552154176 podStartE2EDuration="2m15.552154176s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:01.540155678 +0000 UTC m=+159.231764430" watchObservedRunningTime="2025-12-03 06:30:01.552154176 +0000 UTC m=+159.243762928" Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.563324 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.573554 4818 patch_prober.go:28] interesting pod/router-default-5444994796-fz577 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:30:01 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Dec 03 06:30:01 crc kubenswrapper[4818]: [+]process-running ok Dec 03 06:30:01 crc kubenswrapper[4818]: healthz check failed Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.573604 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fz577" podUID="05f7659c-85e9-4337-8741-9958f12895d4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.641735 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:01 crc kubenswrapper[4818]: E1203 06:30:01.642103 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:02.14208991 +0000 UTC m=+159.833698662 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.695526 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-rl64s" podStartSLOduration=136.695510817 podStartE2EDuration="2m16.695510817s" podCreationTimestamp="2025-12-03 06:27:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:01.695252171 +0000 UTC m=+159.386860923" watchObservedRunningTime="2025-12-03 06:30:01.695510817 +0000 UTC m=+159.387119569" Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.742630 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:01 crc kubenswrapper[4818]: E1203 06:30:01.743135 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:02.24311859 +0000 UTC m=+159.934727342 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.753116 4818 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-fv5lc container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.28:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.753187 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" podUID="d7f12c1b-f504-4155-bd29-bcb6ab57b646" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.28:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.759652 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gt4k9" event={"ID":"b6220eaa-3a5d-4c00-8091-fad092c5c77d","Type":"ContainerStarted","Data":"65926024459e422841a3fac98bf76d59b430238eda6524ccf7e0049cff3a62d9"} Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.771216 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6mbg" event={"ID":"10f43944-25cf-4ae7-8298-82635d81b62e","Type":"ContainerStarted","Data":"784e759b563df778ba8f2a6eebe011559da37b3ea5b1129b36b020be8b52d449"} Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.771272 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6mbg" event={"ID":"10f43944-25cf-4ae7-8298-82635d81b62e","Type":"ContainerStarted","Data":"104b04401cf42ae953d589910e2c1fab9015b03c92dc855a77dd4ca52b785bec"} Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.776266 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" event={"ID":"0122f472-9da5-40c1-aa0a-ab6219543b1d","Type":"ContainerStarted","Data":"f137ed23108ec282aaf3bfe52d37e75932d46c14c517ce4064d4af2b425b10d3"} Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.778490 4818 generic.go:334] "Generic (PLEG): container finished" podID="3dbb4222-9650-4663-95bf-bd2fa4f4e678" containerID="0263d9c2588a655d6b39ec42bbe2ff7bb943474dbf0852a1f5e691daea033530" exitCode=0 Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.778549 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg" event={"ID":"3dbb4222-9650-4663-95bf-bd2fa4f4e678","Type":"ContainerDied","Data":"0263d9c2588a655d6b39ec42bbe2ff7bb943474dbf0852a1f5e691daea033530"} Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.784574 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" podStartSLOduration=136.784556339 podStartE2EDuration="2m16.784556339s" podCreationTimestamp="2025-12-03 06:27:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:01.784080717 +0000 UTC m=+159.475689479" watchObservedRunningTime="2025-12-03 06:30:01.784556339 +0000 UTC m=+159.476165091" Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.791043 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j25z9" event={"ID":"06af2589-0701-4b16-a31c-eb39fe1c8b99","Type":"ContainerStarted","Data":"b7a53494ecbf246862a43f0eab8ecf8b27125ba1fe7f1d24bf90f611e0dda818"} Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.793048 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-l6nvl" event={"ID":"c8c049a4-28db-4ac4-807c-637c3837235a","Type":"ContainerStarted","Data":"a11d0c62068973a3c1bf097abf224727e3354b424d1a2eb7ed2c0151ac67565e"} Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.805777 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" event={"ID":"90705934-7171-4811-85d8-8769a47f7fc7","Type":"ContainerStarted","Data":"b3eaaba6eac17f7c01bd2cf2d49b3b3af681897930b2e984b76f984859490651"} Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.807046 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" event={"ID":"a9ca675e-e374-467c-aa38-da781b84e3dd","Type":"ContainerStarted","Data":"3f2acee4e6e24ee9194475d722b8c484e1edf515e059c1a7971b50acc2768b00"} Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.807094 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" event={"ID":"a9ca675e-e374-467c-aa38-da781b84e3dd","Type":"ContainerStarted","Data":"3e048c8b88c4f340f500dda2b9dc7c011fce00d7900ceacf8f2ba6fa735199e4"} Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.859094 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:01 crc kubenswrapper[4818]: E1203 06:30:01.862502 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:02.362470185 +0000 UTC m=+160.054078937 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.933196 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-fz577" podStartSLOduration=135.933177581 podStartE2EDuration="2m15.933177581s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:01.932869304 +0000 UTC m=+159.624478046" watchObservedRunningTime="2025-12-03 06:30:01.933177581 +0000 UTC m=+159.624786333" Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.943789 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-2sn7n" podStartSLOduration=135.943767654 podStartE2EDuration="2m15.943767654s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:01.862146117 +0000 UTC m=+159.553754869" watchObservedRunningTime="2025-12-03 06:30:01.943767654 +0000 UTC m=+159.635376416" Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.958036 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-2sn7n" Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.963695 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:01 crc kubenswrapper[4818]: E1203 06:30:01.965388 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:02.46533341 +0000 UTC m=+160.156942162 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:01 crc kubenswrapper[4818]: I1203 06:30:01.992236 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gt4k9" podStartSLOduration=135.992200758 podStartE2EDuration="2m15.992200758s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:01.985487921 +0000 UTC m=+159.677096673" watchObservedRunningTime="2025-12-03 06:30:01.992200758 +0000 UTC m=+159.683809510" Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.005498 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l6mbg" podStartSLOduration=136.005480657 podStartE2EDuration="2m16.005480657s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:02.001865258 +0000 UTC m=+159.693474020" watchObservedRunningTime="2025-12-03 06:30:02.005480657 +0000 UTC m=+159.697089419" Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.049746 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-l6nvl" podStartSLOduration=6.049728607 podStartE2EDuration="6.049728607s" podCreationTimestamp="2025-12-03 06:29:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:02.048510146 +0000 UTC m=+159.740118908" watchObservedRunningTime="2025-12-03 06:30:02.049728607 +0000 UTC m=+159.741337359" Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.067590 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:02 crc kubenswrapper[4818]: E1203 06:30:02.067948 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:02.567930269 +0000 UTC m=+160.259539101 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.071419 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l7pnq" podStartSLOduration=137.071403915 podStartE2EDuration="2m17.071403915s" podCreationTimestamp="2025-12-03 06:27:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:02.069531989 +0000 UTC m=+159.761140741" watchObservedRunningTime="2025-12-03 06:30:02.071403915 +0000 UTC m=+159.763012657" Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.073178 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.169018 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:02 crc kubenswrapper[4818]: E1203 06:30:02.169327 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:02.669312897 +0000 UTC m=+160.360921649 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.224782 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" podStartSLOduration=136.224764535 podStartE2EDuration="2m16.224764535s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:02.14846425 +0000 UTC m=+159.840073002" watchObservedRunningTime="2025-12-03 06:30:02.224764535 +0000 UTC m=+159.916373287" Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.270146 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:02 crc kubenswrapper[4818]: E1203 06:30:02.270573 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:02.770557303 +0000 UTC m=+160.462166055 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.371454 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:02 crc kubenswrapper[4818]: E1203 06:30:02.371632 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:02.871604353 +0000 UTC m=+160.563213105 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.372026 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:02 crc kubenswrapper[4818]: E1203 06:30:02.372395 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:02.872382812 +0000 UTC m=+160.563991564 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.423952 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jvbsr"] Dec 03 06:30:02 crc kubenswrapper[4818]: W1203 06:30:02.437929 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d0ddd49_ca6f_44aa_a3aa_5a7717845156.slice/crio-1574593696e1dbd14e5b3631538af10a1f1ad02c346949ea06e9054d2ae824ba WatchSource:0}: Error finding container 1574593696e1dbd14e5b3631538af10a1f1ad02c346949ea06e9054d2ae824ba: Status 404 returned error can't find the container with id 1574593696e1dbd14e5b3631538af10a1f1ad02c346949ea06e9054d2ae824ba Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.448057 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt"] Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.474060 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:02 crc kubenswrapper[4818]: E1203 06:30:02.474637 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:02.974613132 +0000 UTC m=+160.666221884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:02 crc kubenswrapper[4818]: W1203 06:30:02.484531 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaaeda6c3_f0ec_48e5_b7a7_c152ad6c1721.slice/crio-e4a9a74e381d7633bc768f2a4144c66b1ef879f20d95e3a8ef4e3c0bca8a39e5 WatchSource:0}: Error finding container e4a9a74e381d7633bc768f2a4144c66b1ef879f20d95e3a8ef4e3c0bca8a39e5: Status 404 returned error can't find the container with id e4a9a74e381d7633bc768f2a4144c66b1ef879f20d95e3a8ef4e3c0bca8a39e5 Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.494453 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6h7lb"] Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.511726 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-sccrw"] Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.554264 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qnxnn"] Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.561094 4818 patch_prober.go:28] interesting pod/router-default-5444994796-fz577 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:30:02 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Dec 03 06:30:02 crc kubenswrapper[4818]: [+]process-running ok Dec 03 06:30:02 crc kubenswrapper[4818]: healthz check failed Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.561133 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fz577" podUID="05f7659c-85e9-4337-8741-9958f12895d4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.575546 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:02 crc kubenswrapper[4818]: E1203 06:30:02.575867 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:03.075856637 +0000 UTC m=+160.767465389 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.599498 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j8d4t"] Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.645105 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-pvb2q"] Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.676058 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:02 crc kubenswrapper[4818]: E1203 06:30:02.676344 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:03.176328963 +0000 UTC m=+160.867937715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:02 crc kubenswrapper[4818]: W1203 06:30:02.677834 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b90f9f4_fcfa_46b3_80de_b8e335c84e5b.slice/crio-def37116737c42cec530cf0d7914d8c4c1dc4aa0600ba4cd41dece03e0b341db WatchSource:0}: Error finding container def37116737c42cec530cf0d7914d8c4c1dc4aa0600ba4cd41dece03e0b341db: Status 404 returned error can't find the container with id def37116737c42cec530cf0d7914d8c4c1dc4aa0600ba4cd41dece03e0b341db Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.702282 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8x7mv"] Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.791253 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:02 crc kubenswrapper[4818]: E1203 06:30:02.793540 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:03.293519214 +0000 UTC m=+160.985127966 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.798030 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6glnm"] Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.798068 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c"] Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.811583 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-79bwt"] Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.813333 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl"] Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.840899 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dh7z8"] Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.885541 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jvbsr" event={"ID":"9d0ddd49-ca6f-44aa-a3aa-5a7717845156","Type":"ContainerStarted","Data":"1574593696e1dbd14e5b3631538af10a1f1ad02c346949ea06e9054d2ae824ba"} Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.889395 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gkgqw"] Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.893150 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sccrw" event={"ID":"6d2839d4-e624-4d61-9225-894a998e9e9e","Type":"ContainerStarted","Data":"2f41b69b4237893c5f9b1a86b6176f06169b181984f8192f8164e15d728b7d65"} Dec 03 06:30:02 crc kubenswrapper[4818]: E1203 06:30:02.895235 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:03.394687807 +0000 UTC m=+161.086296559 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.894017 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.898028 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:02 crc kubenswrapper[4818]: E1203 06:30:02.898355 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:03.398341768 +0000 UTC m=+161.089950520 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.905522 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j25z9" event={"ID":"06af2589-0701-4b16-a31c-eb39fe1c8b99","Type":"ContainerStarted","Data":"08895c023e1d4f6d3b3b271171fef1b93a9fd4cc06b202cc54e877aa683a7991"} Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.942841 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-pvb2q" event={"ID":"4b90f9f4-fcfa-46b3-80de-b8e335c84e5b","Type":"ContainerStarted","Data":"def37116737c42cec530cf0d7914d8c4c1dc4aa0600ba4cd41dece03e0b341db"} Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.950311 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt" event={"ID":"aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721","Type":"ContainerStarted","Data":"e4a9a74e381d7633bc768f2a4144c66b1ef879f20d95e3a8ef4e3c0bca8a39e5"} Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.953455 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8x7mv" event={"ID":"3deb8606-51f3-4a91-bcfd-0533f315e829","Type":"ContainerStarted","Data":"dd0085626797db26777ae2ed79cbf6c1fc4a1a551612b770da7a3d5fa5a00ccd"} Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.965786 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg" event={"ID":"3dbb4222-9650-4663-95bf-bd2fa4f4e678","Type":"ContainerStarted","Data":"451e22460321e3e34d1b4a36a83c9150e1b5c0930f9e8d630a5f0e3da5d078f4"} Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.966467 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg" Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.968375 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" event={"ID":"39dce7bf-1502-40fa-9d9c-41378ccad702","Type":"ContainerStarted","Data":"e455d47bd5977856c3fa764fa0273fd8406445d19eee8295a0db11ac4cf9ea99"} Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.974350 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" event={"ID":"d167fcbe-c833-43f2-ada0-885857d568a9","Type":"ContainerStarted","Data":"9f2eb0b92b1e893b36928ec7a4b46e6a4ce28eae6525ae843d6c0f81bc0a120e"} Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.986644 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" event={"ID":"a9ca675e-e374-467c-aa38-da781b84e3dd","Type":"ContainerStarted","Data":"d2c6a88d84372767f3a2db429461cb7083d054de13760ae53984c3301360c93a"} Dec 03 06:30:02 crc kubenswrapper[4818]: I1203 06:30:02.998447 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:02 crc kubenswrapper[4818]: E1203 06:30:02.999889 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:03.49986988 +0000 UTC m=+161.191478632 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.075388 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-j25z9" podStartSLOduration=137.075369216 podStartE2EDuration="2m17.075369216s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:03.070566837 +0000 UTC m=+160.762175589" watchObservedRunningTime="2025-12-03 06:30:03.075369216 +0000 UTC m=+160.766977978" Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.087531 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r"] Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.102806 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:03 crc kubenswrapper[4818]: E1203 06:30:03.105227 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:03.605212477 +0000 UTC m=+161.296821439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.120526 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg" podStartSLOduration=137.120499517 podStartE2EDuration="2m17.120499517s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:03.100436319 +0000 UTC m=+160.792045071" watchObservedRunningTime="2025-12-03 06:30:03.120499517 +0000 UTC m=+160.812108269" Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.132944 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-nmqf4"] Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.142110 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-s8vm2"] Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.145387 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv"] Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.152181 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf"] Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.181395 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-596cp"] Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.185068 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cnpz5" podStartSLOduration=137.18504359 podStartE2EDuration="2m17.18504359s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:03.129947212 +0000 UTC m=+160.821555964" watchObservedRunningTime="2025-12-03 06:30:03.18504359 +0000 UTC m=+160.876652342" Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.192934 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ltb69"] Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.197856 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-phwq8"] Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.197885 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9"] Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.198891 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s8mcw"] Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.205721 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz"] Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.207281 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.208270 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4dh8d"] Dec 03 06:30:03 crc kubenswrapper[4818]: E1203 06:30:03.208503 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:03.708476612 +0000 UTC m=+161.400085364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.211172 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:03 crc kubenswrapper[4818]: E1203 06:30:03.211734 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:03.711711053 +0000 UTC m=+161.403319805 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.311810 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:03 crc kubenswrapper[4818]: E1203 06:30:03.312208 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:03.812192239 +0000 UTC m=+161.503800991 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.413927 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:03 crc kubenswrapper[4818]: E1203 06:30:03.414327 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:03.914307616 +0000 UTC m=+161.605916368 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:03 crc kubenswrapper[4818]: W1203 06:30:03.476666 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a2888eb_7da9_4ef2_841b_22f170e66a54.slice/crio-b9bdef36a8644693ae57d8cff818a4f552e0963e5bbe7fd8add584cb1386f21f WatchSource:0}: Error finding container b9bdef36a8644693ae57d8cff818a4f552e0963e5bbe7fd8add584cb1386f21f: Status 404 returned error can't find the container with id b9bdef36a8644693ae57d8cff818a4f552e0963e5bbe7fd8add584cb1386f21f Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.514376 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:03 crc kubenswrapper[4818]: E1203 06:30:03.515029 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:04.014955265 +0000 UTC m=+161.706564017 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.563572 4818 patch_prober.go:28] interesting pod/router-default-5444994796-fz577 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:30:03 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Dec 03 06:30:03 crc kubenswrapper[4818]: [+]process-running ok Dec 03 06:30:03 crc kubenswrapper[4818]: healthz check failed Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.563621 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fz577" podUID="05f7659c-85e9-4337-8741-9958f12895d4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.616268 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:03 crc kubenswrapper[4818]: E1203 06:30:03.616676 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:04.116662632 +0000 UTC m=+161.808271384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.718420 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:03 crc kubenswrapper[4818]: E1203 06:30:03.718708 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:04.218674476 +0000 UTC m=+161.910283378 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.718796 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:03 crc kubenswrapper[4818]: E1203 06:30:03.719288 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:04.219269961 +0000 UTC m=+161.910878713 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.820145 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:03 crc kubenswrapper[4818]: E1203 06:30:03.820695 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:04.32067176 +0000 UTC m=+162.012280502 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.922535 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:03 crc kubenswrapper[4818]: E1203 06:30:03.923308 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:04.423269108 +0000 UTC m=+162.114877880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.983099 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.983153 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:30:03 crc kubenswrapper[4818]: I1203 06:30:03.994327 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.020248 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s8mcw" event={"ID":"8935f198-82ab-463c-96d5-300a23d613e6","Type":"ContainerStarted","Data":"66e9863381aa731dfcdcf80cb28d250f7cfdd58cf7175414be4d98e7251278f6"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.023534 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:04 crc kubenswrapper[4818]: E1203 06:30:04.024024 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:04.523995591 +0000 UTC m=+162.215604343 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.046527 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" event={"ID":"2bf3947e-3c18-4ef1-aca9-98419c558d17","Type":"ContainerStarted","Data":"72e76f89c31ecc7373dc1bd8f091bfb859eb6023937bd1f1f76f08414ab2a848"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.050431 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf" event={"ID":"503c8137-bff7-476c-90f1-6c0d8112cc2e","Type":"ContainerStarted","Data":"ab1c6c81acf407ec6339eae5e7ac47f95849b703e7ff1d482607d577b0fabe97"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.051328 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.052487 4818 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-dr7rf container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.052535 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf" podUID="503c8137-bff7-476c-90f1-6c0d8112cc2e" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.064144 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" event={"ID":"b14853e5-2388-4dba-a3ca-dd9453ea3452","Type":"ContainerStarted","Data":"168f9746ddb0f9f01cd5fa8671da826383f6291d2b9c1d1db66831ff1cd54e07"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.069495 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv" event={"ID":"1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6","Type":"ContainerStarted","Data":"60641d0197250843e971b25c6b7252262de40d1fea3a57de652955106bbacfd6"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.078515 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf" podStartSLOduration=138.078495885 podStartE2EDuration="2m18.078495885s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:04.076437184 +0000 UTC m=+161.768045946" watchObservedRunningTime="2025-12-03 06:30:04.078495885 +0000 UTC m=+161.770104637" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.089202 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-pvb2q" event={"ID":"4b90f9f4-fcfa-46b3-80de-b8e335c84e5b","Type":"ContainerStarted","Data":"4b150bedf3182047ef0e73f6950cba9a1da762611ed6f61fa3cf2f3970d4c321"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.089967 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-pvb2q" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.093986 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-pvb2q container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.094041 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-pvb2q" podUID="4b90f9f4-fcfa-46b3-80de-b8e335c84e5b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.101233 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sccrw" event={"ID":"6d2839d4-e624-4d61-9225-894a998e9e9e","Type":"ContainerStarted","Data":"81e079fab9ada721df205aef6940ff13656d56bcf1aebbc1b718d9e56c04f368"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.108596 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6glnm" event={"ID":"a3eedabc-4a51-4068-a9da-c8a867cb3219","Type":"ContainerStarted","Data":"d0b494272a979ea61738df2453367c47c768e9b58ffe8dc5b4cf0e8d2159d4f3"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.108648 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6glnm" event={"ID":"a3eedabc-4a51-4068-a9da-c8a867cb3219","Type":"ContainerStarted","Data":"5d066158db5bb7c501bfb1e16b65d5b456ee67b4e5aadab63b15f63ebffbbafe"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.108667 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6glnm" event={"ID":"a3eedabc-4a51-4068-a9da-c8a867cb3219","Type":"ContainerStarted","Data":"6707edbb93e02d9a01f8ca38c01d5cce777072b6f9b5f793a8a5da7d4fd56025"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.109377 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6glnm" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.121312 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" event={"ID":"d167fcbe-c833-43f2-ada0-885857d568a9","Type":"ContainerStarted","Data":"1026dfc709b232473f6ded1bdc37d4baafb513f108710ba4931a17eb79d3af39"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.124632 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:04 crc kubenswrapper[4818]: E1203 06:30:04.125240 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:04.625222275 +0000 UTC m=+162.316831027 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.125473 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" event={"ID":"f033be5e-5c71-4aeb-8825-bcf234d174a1","Type":"ContainerStarted","Data":"dbd2e588851182739b2867e2911ed72b1374191776428f862721d1bc397c4dde"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.127317 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j8d4t" event={"ID":"dae2df0a-0703-40b5-9040-04b3af885585","Type":"ContainerStarted","Data":"932c98367cc27181471a07f663986040a8272d0fef77d5a1267c9aaf6b0b0c57"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.127427 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j8d4t" event={"ID":"dae2df0a-0703-40b5-9040-04b3af885585","Type":"ContainerStarted","Data":"d5fd347625d0722d2fe005d3c99f558eff2033724efecfb4c1731694add1f084"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.147142 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-sccrw" podStartSLOduration=138.147115609 podStartE2EDuration="2m18.147115609s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:04.146680578 +0000 UTC m=+161.838289330" watchObservedRunningTime="2025-12-03 06:30:04.147115609 +0000 UTC m=+161.838724361" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.148275 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-pvb2q" podStartSLOduration=138.148261728 podStartE2EDuration="2m18.148261728s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:04.112411587 +0000 UTC m=+161.804020339" watchObservedRunningTime="2025-12-03 06:30:04.148261728 +0000 UTC m=+161.839870480" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.150415 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt" event={"ID":"aaeda6c3-f0ec-48e5-b7a7-c152ad6c1721","Type":"ContainerStarted","Data":"46e672852f7bfcd9203600590105f9bb0c4adde61e5d387f1ec07175f125a331"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.160358 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-596cp" event={"ID":"0ff5cea8-7b98-4a15-87cb-f07fb9cfabea","Type":"ContainerStarted","Data":"7b31bf4d3cc86945f87ee69c295d79e03fe5799b98227fb2cc9f1acbac813307"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.166027 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6glnm" podStartSLOduration=138.165996528 podStartE2EDuration="2m18.165996528s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:04.163937637 +0000 UTC m=+161.855546389" watchObservedRunningTime="2025-12-03 06:30:04.165996528 +0000 UTC m=+161.857605280" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.170212 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" event={"ID":"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7","Type":"ContainerStarted","Data":"4aed2b700618ebe91a04599926c9a808571c294cd62fcc3227d8b407615af53f"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.170486 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" event={"ID":"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7","Type":"ContainerStarted","Data":"a6f7496ab46ed22b151f3c734f177cf43ec75578d8c47f658548896a5221ccac"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.170451 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" podUID="1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7" containerName="collect-profiles" containerID="cri-o://4aed2b700618ebe91a04599926c9a808571c294cd62fcc3227d8b407615af53f" gracePeriod=30 Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.185181 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dh7z8" event={"ID":"2c313309-b494-4209-a248-cb526aa88f36","Type":"ContainerStarted","Data":"cc5d36448e205133bd2a6a334b0a7ba8e5bc22eb7a8fa0f4a0b56dd328a1b68f"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.188353 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" event={"ID":"0c633e2b-4169-4277-a64b-85c249869b0c","Type":"ContainerStarted","Data":"676e202907e2cfaa97637bc5d9e957f0acc4eab1744d8dea8fb9bfe5405c19f5"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.188491 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" event={"ID":"0c633e2b-4169-4277-a64b-85c249869b0c","Type":"ContainerStarted","Data":"108a5accb8609fdd8c0bcc8669dd432e0c7e78c2c700dbf107e4a46fa6b4a503"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.188967 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.189650 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jvbsr" event={"ID":"9d0ddd49-ca6f-44aa-a3aa-5a7717845156","Type":"ContainerStarted","Data":"048ecd15843dfb515cbb86deb15d13131a5512b43246c9a8a8e92fc8ce1fc9bc"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.189936 4818 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-f55bl container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.190053 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" podUID="0c633e2b-4169-4277-a64b-85c249869b0c" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.195734 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k4pkt" podStartSLOduration=138.195706056 podStartE2EDuration="2m18.195706056s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:04.192655471 +0000 UTC m=+161.884264223" watchObservedRunningTime="2025-12-03 06:30:04.195706056 +0000 UTC m=+161.887314808" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.199686 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-79bwt" event={"ID":"3a63ab77-ab0e-4f43-9d21-2ec684236ce9","Type":"ContainerStarted","Data":"c0ca276787edd81a815ee44a1c550f014a29f78aca7fc9a7fd572f0aa6650507"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.199737 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-79bwt" event={"ID":"3a63ab77-ab0e-4f43-9d21-2ec684236ce9","Type":"ContainerStarted","Data":"45ce5a2a4f665b3bce29eb791720e8a36c5ca99b9f6eb2577a2d8f572a913ffe"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.201171 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ltb69" event={"ID":"faad9125-f494-49ae-ba77-941e6abe67b9","Type":"ContainerStarted","Data":"4ce9d4f723bbfa4998b27ecc261c6f37c4e438d6ec51535ede7181a12a3d10f1"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.205022 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8x7mv" event={"ID":"3deb8606-51f3-4a91-bcfd-0533f315e829","Type":"ContainerStarted","Data":"79fa4a5e373b6563b50735bddac0856d3b8feb0225766aaf3f416c08aa5c6907"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.207718 4818 generic.go:334] "Generic (PLEG): container finished" podID="39dce7bf-1502-40fa-9d9c-41378ccad702" containerID="0a04b494d6f3d9a9da03ef50d43bef6ddd58e8ff82fc4d2ccc1ac21a872ba6a4" exitCode=0 Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.208493 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" event={"ID":"39dce7bf-1502-40fa-9d9c-41378ccad702","Type":"ContainerDied","Data":"0a04b494d6f3d9a9da03ef50d43bef6ddd58e8ff82fc4d2ccc1ac21a872ba6a4"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.219216 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-nmqf4" event={"ID":"a0dc41b5-d81d-4a47-aaad-8a9d8216e513","Type":"ContainerStarted","Data":"b29eb7837ac02dd8394c5ce78d39fa36dac4cfc2f733fed9b3c89f4970d0b6bb"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.225258 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-6h7lb" podStartSLOduration=138.22524276 podStartE2EDuration="2m18.22524276s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:04.2236232 +0000 UTC m=+161.915231952" watchObservedRunningTime="2025-12-03 06:30:04.22524276 +0000 UTC m=+161.916851512" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.227767 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.228374 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" event={"ID":"53a413a7-34f8-4fb3-893b-c4fa9711017a","Type":"ContainerStarted","Data":"03df0cf759afe73c38c7041c610ba37cf8b523d7b5ed1ac477c6aa9c1a7c38f7"} Dec 03 06:30:04 crc kubenswrapper[4818]: E1203 06:30:04.229041 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:04.729026924 +0000 UTC m=+162.420635676 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.229417 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.242805 4818 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4dh8d container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.242870 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" podUID="53a413a7-34f8-4fb3-893b-c4fa9711017a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.273677 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c" event={"ID":"0aef8214-0bed-45cc-b3cd-77fcfbecb7aa","Type":"ContainerStarted","Data":"1930757cf69222fb969b73faf6140af185ec630ff5f46d2f88e1b5b2bd3571e8"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.273724 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c" event={"ID":"0aef8214-0bed-45cc-b3cd-77fcfbecb7aa","Type":"ContainerStarted","Data":"c6869d93b1c6ea9148b0527cb1d9a7222343ab8781fb92b1c3d88ae7c74a7146"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.274182 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.277952 4818 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-jr26c container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.278016 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c" podUID="0aef8214-0bed-45cc-b3cd-77fcfbecb7aa" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.282084 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-s8vm2" event={"ID":"8b9018a9-c76c-4471-8b27-06c1cbc065eb","Type":"ContainerStarted","Data":"00806ddf5cd496c086783fd51b3d69645e15b1e1242a2c2ef9ad4dbe53ffa25e"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.293961 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-phwq8" event={"ID":"9a2888eb-7da9-4ef2-841b-22f170e66a54","Type":"ContainerStarted","Data":"b9bdef36a8644693ae57d8cff818a4f552e0963e5bbe7fd8add584cb1386f21f"} Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.304541 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-npq7z" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.305231 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dh7z8" podStartSLOduration=138.305223677 podStartE2EDuration="2m18.305223677s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:04.303438693 +0000 UTC m=+161.995047445" watchObservedRunningTime="2025-12-03 06:30:04.305223677 +0000 UTC m=+161.996832429" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.329266 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:04 crc kubenswrapper[4818]: E1203 06:30:04.331954 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:04.831937091 +0000 UTC m=+162.523545923 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.368147 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" podStartSLOduration=139.3681298 podStartE2EDuration="2m19.3681298s" podCreationTimestamp="2025-12-03 06:27:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:04.365612347 +0000 UTC m=+162.057221099" watchObservedRunningTime="2025-12-03 06:30:04.3681298 +0000 UTC m=+162.059738552" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.418198 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" podStartSLOduration=138.418179813 podStartE2EDuration="2m18.418179813s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:04.416256046 +0000 UTC m=+162.107864798" watchObservedRunningTime="2025-12-03 06:30:04.418179813 +0000 UTC m=+162.109788565" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.429727 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:04 crc kubenswrapper[4818]: E1203 06:30:04.431210 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:04.931175526 +0000 UTC m=+162.622784278 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.435495 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:04 crc kubenswrapper[4818]: E1203 06:30:04.435876 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:04.935864283 +0000 UTC m=+162.627473025 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.548562 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:04 crc kubenswrapper[4818]: E1203 06:30:04.549294 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:05.04927515 +0000 UTC m=+162.740883902 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.559150 4818 patch_prober.go:28] interesting pod/router-default-5444994796-fz577 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:30:04 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Dec 03 06:30:04 crc kubenswrapper[4818]: [+]process-running ok Dec 03 06:30:04 crc kubenswrapper[4818]: healthz check failed Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.559207 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fz577" podUID="05f7659c-85e9-4337-8741-9958f12895d4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.601760 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c" podStartSLOduration=138.601739504 podStartE2EDuration="2m18.601739504s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:04.600188165 +0000 UTC m=+162.291796917" watchObservedRunningTime="2025-12-03 06:30:04.601739504 +0000 UTC m=+162.293348256" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.602192 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" podStartSLOduration=138.602186665 podStartE2EDuration="2m18.602186665s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:04.509095062 +0000 UTC m=+162.200703814" watchObservedRunningTime="2025-12-03 06:30:04.602186665 +0000 UTC m=+162.293795417" Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.653514 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:04 crc kubenswrapper[4818]: E1203 06:30:04.653849 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:05.153835568 +0000 UTC m=+162.845444320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.756238 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:04 crc kubenswrapper[4818]: E1203 06:30:04.757099 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:05.257073752 +0000 UTC m=+162.948682504 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.757492 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:04 crc kubenswrapper[4818]: E1203 06:30:04.758183 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:05.258165499 +0000 UTC m=+162.949774251 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.858890 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:04 crc kubenswrapper[4818]: E1203 06:30:04.859341 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:05.359322982 +0000 UTC m=+163.050931734 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:04 crc kubenswrapper[4818]: I1203 06:30:04.968643 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:04 crc kubenswrapper[4818]: E1203 06:30:04.969088 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:05.469069449 +0000 UTC m=+163.160678201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.026367 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_collect-profiles-29412375-f596r_1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7/collect-profiles/0.log" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.026452 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.071186 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:05 crc kubenswrapper[4818]: E1203 06:30:05.071690 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:05.571671597 +0000 UTC m=+163.263280339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.172378 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flzcx\" (UniqueName: \"kubernetes.io/projected/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-kube-api-access-flzcx\") pod \"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7\" (UID: \"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7\") " Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.173000 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-config-volume\") pod \"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7\" (UID: \"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7\") " Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.173097 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-secret-volume\") pod \"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7\" (UID: \"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7\") " Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.173319 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:05 crc kubenswrapper[4818]: E1203 06:30:05.173630 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:05.67361375 +0000 UTC m=+163.365222512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.181194 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-config-volume" (OuterVolumeSpecName: "config-volume") pod "1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7" (UID: "1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.181882 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-kube-api-access-flzcx" (OuterVolumeSpecName: "kube-api-access-flzcx") pod "1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7" (UID: "1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7"). InnerVolumeSpecName "kube-api-access-flzcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.203346 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7" (UID: "1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.277352 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:05 crc kubenswrapper[4818]: E1203 06:30:05.277543 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:05.777518451 +0000 UTC m=+163.469127203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.277763 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.277875 4818 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.277892 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flzcx\" (UniqueName: \"kubernetes.io/projected/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-kube-api-access-flzcx\") on node \"crc\" DevicePath \"\"" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.277902 4818 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 06:30:05 crc kubenswrapper[4818]: E1203 06:30:05.278180 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:05.778172357 +0000 UTC m=+163.469781109 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.305380 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" event={"ID":"f033be5e-5c71-4aeb-8825-bcf234d174a1","Type":"ContainerStarted","Data":"820336f7aa4564a8c9d5561a937f63298b3615bc679942f01a71a37b9a3328be"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.309275 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-s8vm2" event={"ID":"8b9018a9-c76c-4471-8b27-06c1cbc065eb","Type":"ContainerStarted","Data":"ddab1f929f6d000c3c828889207c97c5d28f220f53da2fe0ce16545f24a4e29f"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.309301 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-s8vm2" event={"ID":"8b9018a9-c76c-4471-8b27-06c1cbc065eb","Type":"ContainerStarted","Data":"6c767efe8209bd013e5c52ba62e0520123bb0730f865457f0b3280265f5d35b8"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.311554 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jvbsr" event={"ID":"9d0ddd49-ca6f-44aa-a3aa-5a7717845156","Type":"ContainerStarted","Data":"1b5888bd7351517946d6858dd4ff92393a39d5dae18b16b68893fc08e162405f"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.315415 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-596cp" event={"ID":"0ff5cea8-7b98-4a15-87cb-f07fb9cfabea","Type":"ContainerStarted","Data":"181d3a6a530e1e85ff9c4a6a88641f1ffbb4634528c07203f3140ae8c3faab70"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.320725 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" event={"ID":"b14853e5-2388-4dba-a3ca-dd9453ea3452","Type":"ContainerStarted","Data":"4b4401e3309de7862473306a82d73335ab0fc91ab95b413a1e82920ee9c2fc52"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.322849 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-nmqf4" event={"ID":"a0dc41b5-d81d-4a47-aaad-8a9d8216e513","Type":"ContainerStarted","Data":"3a81d2efbffe540f8022d384c7911db4d177bf8a6dbc6f9a0e82008da132d9da"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.330486 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" podStartSLOduration=5.330464746 podStartE2EDuration="5.330464746s" podCreationTimestamp="2025-12-03 06:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:05.329327248 +0000 UTC m=+163.020936000" watchObservedRunningTime="2025-12-03 06:30:05.330464746 +0000 UTC m=+163.022073498" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.332372 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv" event={"ID":"1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6","Type":"ContainerStarted","Data":"0fbfc134419416453c73001e866b5b9758c705b17a72e32365efd2b73311096b"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.332412 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv" event={"ID":"1f6aeec1-ce4c-4a85-a188-a0bb94f79bb6","Type":"ContainerStarted","Data":"904eed0df584134bff4d0771943568331646d8110537136d0b6a2f98520c082b"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.352991 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-79bwt" event={"ID":"3a63ab77-ab0e-4f43-9d21-2ec684236ce9","Type":"ContainerStarted","Data":"b7aae128a28b28f5c251289972c1b203ba413f6f833b04e9d6cbd21bb63a9063"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.353999 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-79bwt" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.361893 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-nmqf4" podStartSLOduration=9.361881067 podStartE2EDuration="9.361881067s" podCreationTimestamp="2025-12-03 06:29:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:05.361160639 +0000 UTC m=+163.052769391" watchObservedRunningTime="2025-12-03 06:30:05.361881067 +0000 UTC m=+163.053489809" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.375670 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" event={"ID":"53a413a7-34f8-4fb3-893b-c4fa9711017a","Type":"ContainerStarted","Data":"dddd0b228babbc5d22cd3ceeca18a47a59bd31fc0b34de7805a71b039c0c2fcb"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.379907 4818 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4dh8d container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.380038 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" podUID="53a413a7-34f8-4fb3-893b-c4fa9711017a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.381371 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:05 crc kubenswrapper[4818]: E1203 06:30:05.381720 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:05.881705529 +0000 UTC m=+163.573314281 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.390122 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s8mcw" event={"ID":"8935f198-82ab-463c-96d5-300a23d613e6","Type":"ContainerStarted","Data":"bcbffffa71188b459f7a6a1bfd5e04e6f2bb00d35de19087cc6c66cb0a2a8cd4"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.390565 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-jvbsr" podStartSLOduration=139.390549229 podStartE2EDuration="2m19.390549229s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:05.389310328 +0000 UTC m=+163.080919070" watchObservedRunningTime="2025-12-03 06:30:05.390549229 +0000 UTC m=+163.082157981" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.405076 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_collect-profiles-29412375-f596r_1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7/collect-profiles/0.log" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.405129 4818 generic.go:334] "Generic (PLEG): container finished" podID="1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7" containerID="4aed2b700618ebe91a04599926c9a808571c294cd62fcc3227d8b407615af53f" exitCode=2 Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.405214 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" event={"ID":"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7","Type":"ContainerDied","Data":"4aed2b700618ebe91a04599926c9a808571c294cd62fcc3227d8b407615af53f"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.405246 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" event={"ID":"1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7","Type":"ContainerDied","Data":"a6f7496ab46ed22b151f3c734f177cf43ec75578d8c47f658548896a5221ccac"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.405265 4818 scope.go:117] "RemoveContainer" containerID="4aed2b700618ebe91a04599926c9a808571c294cd62fcc3227d8b407615af53f" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.405424 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.414740 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-596cp" podStartSLOduration=139.414721 podStartE2EDuration="2m19.414721s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:05.411867449 +0000 UTC m=+163.103476201" watchObservedRunningTime="2025-12-03 06:30:05.414721 +0000 UTC m=+163.106329752" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.432232 4818 scope.go:117] "RemoveContainer" containerID="4aed2b700618ebe91a04599926c9a808571c294cd62fcc3227d8b407615af53f" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.440759 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dh7z8" event={"ID":"2c313309-b494-4209-a248-cb526aa88f36","Type":"ContainerStarted","Data":"4e6c6ab70d3d206ada0e2e5b8c22087e210efe1c9c8be4a62d2c7d2b0d2a58a8"} Dec 03 06:30:05 crc kubenswrapper[4818]: E1203 06:30:05.441131 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aed2b700618ebe91a04599926c9a808571c294cd62fcc3227d8b407615af53f\": container with ID starting with 4aed2b700618ebe91a04599926c9a808571c294cd62fcc3227d8b407615af53f not found: ID does not exist" containerID="4aed2b700618ebe91a04599926c9a808571c294cd62fcc3227d8b407615af53f" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.441522 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aed2b700618ebe91a04599926c9a808571c294cd62fcc3227d8b407615af53f"} err="failed to get container status \"4aed2b700618ebe91a04599926c9a808571c294cd62fcc3227d8b407615af53f\": rpc error: code = NotFound desc = could not find container \"4aed2b700618ebe91a04599926c9a808571c294cd62fcc3227d8b407615af53f\": container with ID starting with 4aed2b700618ebe91a04599926c9a808571c294cd62fcc3227d8b407615af53f not found: ID does not exist" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.447624 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ltb69" event={"ID":"faad9125-f494-49ae-ba77-941e6abe67b9","Type":"ContainerStarted","Data":"06a3343d811ef63969ff40cbd92265ab204bec8a16e567090c05a9a071511f15"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.464672 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" event={"ID":"39dce7bf-1502-40fa-9d9c-41378ccad702","Type":"ContainerStarted","Data":"fbaed7766b9c6c4b772b9b5b357257535087dfb1ce9d2e658bee568739839685"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.471909 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-s8vm2" podStartSLOduration=139.47189197 podStartE2EDuration="2m19.47189197s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:05.47150161 +0000 UTC m=+163.163110362" watchObservedRunningTime="2025-12-03 06:30:05.47189197 +0000 UTC m=+163.163500722" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.479032 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h6fkg" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.482722 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.484453 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-phwq8" event={"ID":"9a2888eb-7da9-4ef2-841b-22f170e66a54","Type":"ContainerStarted","Data":"a32b1526d6cf7a36f04a452aa3335992b032e6ec544f03045baba2a86f019dc6"} Dec 03 06:30:05 crc kubenswrapper[4818]: E1203 06:30:05.484556 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:05.984540584 +0000 UTC m=+163.676149336 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.503301 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf" event={"ID":"503c8137-bff7-476c-90f1-6c0d8112cc2e","Type":"ContainerStarted","Data":"4d658172753f6795b8e946b03ae4c655f02d2192bc70c65c406797b86d079964"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.508971 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ltb69" podStartSLOduration=139.50895351 podStartE2EDuration="2m19.50895351s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:05.50852661 +0000 UTC m=+163.200135372" watchObservedRunningTime="2025-12-03 06:30:05.50895351 +0000 UTC m=+163.200562262" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.520448 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dr7rf" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.520493 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" event={"ID":"2bf3947e-3c18-4ef1-aca9-98419c558d17","Type":"ContainerStarted","Data":"a9363bc5282716fdc1ba31fd8f3ba772f9525ec64a070f21e4f67c49389e33fa"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.520513 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" event={"ID":"2bf3947e-3c18-4ef1-aca9-98419c558d17","Type":"ContainerStarted","Data":"5c06ce5b4501a56e6fe1c6e593f1a0c76204bd87d1db9c62efcef81cdb6353df"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.523351 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8x7mv" event={"ID":"3deb8606-51f3-4a91-bcfd-0533f315e829","Type":"ContainerStarted","Data":"9318b0000fe25bc5bb34efd5a486a2468a04d7c99ae6371bf39f936e29888df4"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.526915 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j8d4t" event={"ID":"dae2df0a-0703-40b5-9040-04b3af885585","Type":"ContainerStarted","Data":"814bf3d24f1aa3a1a34b018da32eb286b52bae1a493f3c255f20c72d3c605d57"} Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.529062 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-pvb2q container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.529111 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-pvb2q" podUID="4b90f9f4-fcfa-46b3-80de-b8e335c84e5b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.544139 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jr26c" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.550018 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-f55bl" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.560859 4818 patch_prober.go:28] interesting pod/router-default-5444994796-fz577 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:30:05 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Dec 03 06:30:05 crc kubenswrapper[4818]: [+]process-running ok Dec 03 06:30:05 crc kubenswrapper[4818]: healthz check failed Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.560917 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fz577" podUID="05f7659c-85e9-4337-8741-9958f12895d4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.583679 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:05 crc kubenswrapper[4818]: E1203 06:30:05.585295 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:06.085268256 +0000 UTC m=+163.776877048 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.605875 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s8mcw" podStartSLOduration=139.605853078 podStartE2EDuration="2m19.605853078s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:05.558892871 +0000 UTC m=+163.250501633" watchObservedRunningTime="2025-12-03 06:30:05.605853078 +0000 UTC m=+163.297461830" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.624719 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-79bwt" podStartSLOduration=9.624702286 podStartE2EDuration="9.624702286s" podCreationTimestamp="2025-12-03 06:29:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:05.605162681 +0000 UTC m=+163.296771443" watchObservedRunningTime="2025-12-03 06:30:05.624702286 +0000 UTC m=+163.316311038" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.626730 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r"] Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.638893 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412375-f596r"] Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.660354 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-bv9gv" podStartSLOduration=139.660333111 podStartE2EDuration="2m19.660333111s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:05.658645529 +0000 UTC m=+163.350254281" watchObservedRunningTime="2025-12-03 06:30:05.660333111 +0000 UTC m=+163.351941863" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.686380 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:05 crc kubenswrapper[4818]: E1203 06:30:05.697938 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:06.197900514 +0000 UTC m=+163.889509346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.721242 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8x7mv" podStartSLOduration=139.721222944 podStartE2EDuration="2m19.721222944s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:05.68200494 +0000 UTC m=+163.373613692" watchObservedRunningTime="2025-12-03 06:30:05.721222944 +0000 UTC m=+163.412831696" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.757592 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-phwq8" podStartSLOduration=139.757573877 podStartE2EDuration="2m19.757573877s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:05.7552814 +0000 UTC m=+163.446890162" watchObservedRunningTime="2025-12-03 06:30:05.757573877 +0000 UTC m=+163.449182629" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.788169 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:05 crc kubenswrapper[4818]: E1203 06:30:05.788703 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:06.288680889 +0000 UTC m=+163.980289651 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.797090 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8jzsz" podStartSLOduration=139.797079338 podStartE2EDuration="2m19.797079338s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:05.795212052 +0000 UTC m=+163.486820804" watchObservedRunningTime="2025-12-03 06:30:05.797079338 +0000 UTC m=+163.488688090" Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.892595 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:05 crc kubenswrapper[4818]: E1203 06:30:05.893377 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:06.39336404 +0000 UTC m=+164.084972782 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:05 crc kubenswrapper[4818]: I1203 06:30:05.994235 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:05 crc kubenswrapper[4818]: E1203 06:30:05.994524 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:06.494510703 +0000 UTC m=+164.186119455 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.095581 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:06 crc kubenswrapper[4818]: E1203 06:30:06.095938 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:06.595927062 +0000 UTC m=+164.287535814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.197671 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:06 crc kubenswrapper[4818]: E1203 06:30:06.198027 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:06.698012168 +0000 UTC m=+164.389620910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.252577 4818 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.299578 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:06 crc kubenswrapper[4818]: E1203 06:30:06.300046 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:06.800026812 +0000 UTC m=+164.491635624 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.400412 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:06 crc kubenswrapper[4818]: E1203 06:30:06.400624 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:06.900597471 +0000 UTC m=+164.592206223 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.400735 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:06 crc kubenswrapper[4818]: E1203 06:30:06.401159 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:06.901143724 +0000 UTC m=+164.592752506 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.463993 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-j8d4t" podStartSLOduration=140.463967985 podStartE2EDuration="2m20.463967985s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:05.931713283 +0000 UTC m=+163.623322025" watchObservedRunningTime="2025-12-03 06:30:06.463967985 +0000 UTC m=+164.155576737" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.466903 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k7wwb"] Dec 03 06:30:06 crc kubenswrapper[4818]: E1203 06:30:06.467169 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7" containerName="collect-profiles" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.467189 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7" containerName="collect-profiles" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.467356 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7" containerName="collect-profiles" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.468237 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k7wwb" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.474397 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.483878 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k7wwb"] Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.502546 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:06 crc kubenswrapper[4818]: E1203 06:30:06.502776 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:07.002733538 +0000 UTC m=+164.694342290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.502933 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:06 crc kubenswrapper[4818]: E1203 06:30:06.503361 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:07.003344793 +0000 UTC m=+164.694953535 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.534565 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" event={"ID":"39dce7bf-1502-40fa-9d9c-41378ccad702","Type":"ContainerStarted","Data":"8ee912cf751a2f5be89811f18d95609b7bfcb13e8118b4b576d6a610f1a14c8b"} Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.537203 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" event={"ID":"b14853e5-2388-4dba-a3ca-dd9453ea3452","Type":"ContainerStarted","Data":"313b0f015d9d33656602307011fcc4ae772e6d99d5e0b694f40f740729cdd5ec"} Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.537251 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" event={"ID":"b14853e5-2388-4dba-a3ca-dd9453ea3452","Type":"ContainerStarted","Data":"7ae69e897f80ec5a1d9967943b846fe9c47211ea28b6aff01a47ce046a521e9e"} Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.539212 4818 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4dh8d container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.539256 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" podUID="53a413a7-34f8-4fb3-893b-c4fa9711017a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.555490 4818 patch_prober.go:28] interesting pod/router-default-5444994796-fz577 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:30:06 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Dec 03 06:30:06 crc kubenswrapper[4818]: [+]process-running ok Dec 03 06:30:06 crc kubenswrapper[4818]: healthz check failed Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.555969 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fz577" podUID="05f7659c-85e9-4337-8741-9958f12895d4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.573972 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" podStartSLOduration=141.573949007 podStartE2EDuration="2m21.573949007s" podCreationTimestamp="2025-12-03 06:27:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:06.569293951 +0000 UTC m=+164.260902713" watchObservedRunningTime="2025-12-03 06:30:06.573949007 +0000 UTC m=+164.265557779" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.603942 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:06 crc kubenswrapper[4818]: E1203 06:30:06.604245 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:07.104222629 +0000 UTC m=+164.795831391 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.604410 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-catalog-content\") pod \"community-operators-k7wwb\" (UID: \"0301f892-ca1d-4f8b-a429-6d0ad96a50dc\") " pod="openshift-marketplace/community-operators-k7wwb" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.604515 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-utilities\") pod \"community-operators-k7wwb\" (UID: \"0301f892-ca1d-4f8b-a429-6d0ad96a50dc\") " pod="openshift-marketplace/community-operators-k7wwb" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.604615 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tscsg\" (UniqueName: \"kubernetes.io/projected/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-kube-api-access-tscsg\") pod \"community-operators-k7wwb\" (UID: \"0301f892-ca1d-4f8b-a429-6d0ad96a50dc\") " pod="openshift-marketplace/community-operators-k7wwb" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.657081 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jrfsv"] Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.658198 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jrfsv" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.661715 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.668083 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jrfsv"] Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.705889 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tscsg\" (UniqueName: \"kubernetes.io/projected/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-kube-api-access-tscsg\") pod \"community-operators-k7wwb\" (UID: \"0301f892-ca1d-4f8b-a429-6d0ad96a50dc\") " pod="openshift-marketplace/community-operators-k7wwb" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.706779 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.707017 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-catalog-content\") pod \"community-operators-k7wwb\" (UID: \"0301f892-ca1d-4f8b-a429-6d0ad96a50dc\") " pod="openshift-marketplace/community-operators-k7wwb" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.707209 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-utilities\") pod \"community-operators-k7wwb\" (UID: \"0301f892-ca1d-4f8b-a429-6d0ad96a50dc\") " pod="openshift-marketplace/community-operators-k7wwb" Dec 03 06:30:06 crc kubenswrapper[4818]: E1203 06:30:06.710539 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:07.21052355 +0000 UTC m=+164.902132372 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.711723 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-utilities\") pod \"community-operators-k7wwb\" (UID: \"0301f892-ca1d-4f8b-a429-6d0ad96a50dc\") " pod="openshift-marketplace/community-operators-k7wwb" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.717325 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-catalog-content\") pod \"community-operators-k7wwb\" (UID: \"0301f892-ca1d-4f8b-a429-6d0ad96a50dc\") " pod="openshift-marketplace/community-operators-k7wwb" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.745918 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tscsg\" (UniqueName: \"kubernetes.io/projected/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-kube-api-access-tscsg\") pod \"community-operators-k7wwb\" (UID: \"0301f892-ca1d-4f8b-a429-6d0ad96a50dc\") " pod="openshift-marketplace/community-operators-k7wwb" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.755435 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7" path="/var/lib/kubelet/pods/1ef0db87-733c-4c8d-a0ef-dcbf3b0baac7/volumes" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.784999 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k7wwb" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.808441 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:06 crc kubenswrapper[4818]: E1203 06:30:06.808661 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:07.308623607 +0000 UTC m=+165.000232379 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.808842 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c4ffde8-459a-4634-b567-5f900d40ceec-catalog-content\") pod \"certified-operators-jrfsv\" (UID: \"0c4ffde8-459a-4634-b567-5f900d40ceec\") " pod="openshift-marketplace/certified-operators-jrfsv" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.808909 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-879pq\" (UniqueName: \"kubernetes.io/projected/0c4ffde8-459a-4634-b567-5f900d40ceec-kube-api-access-879pq\") pod \"certified-operators-jrfsv\" (UID: \"0c4ffde8-459a-4634-b567-5f900d40ceec\") " pod="openshift-marketplace/certified-operators-jrfsv" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.809004 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c4ffde8-459a-4634-b567-5f900d40ceec-utilities\") pod \"certified-operators-jrfsv\" (UID: \"0c4ffde8-459a-4634-b567-5f900d40ceec\") " pod="openshift-marketplace/certified-operators-jrfsv" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.809050 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:06 crc kubenswrapper[4818]: E1203 06:30:06.809441 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:07.309426667 +0000 UTC m=+165.001035419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.838887 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.839598 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.845878 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.849581 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.860043 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.892845 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zfvqs"] Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.909982 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfvqs" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.910904 4818 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-03T06:30:06.252610354Z","Handler":null,"Name":""} Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.911714 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:06 crc kubenswrapper[4818]: E1203 06:30:06.912455 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:07.412433306 +0000 UTC m=+165.104042058 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.914129 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c4ffde8-459a-4634-b567-5f900d40ceec-utilities\") pod \"certified-operators-jrfsv\" (UID: \"0c4ffde8-459a-4634-b567-5f900d40ceec\") " pod="openshift-marketplace/certified-operators-jrfsv" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.914176 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.914277 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c4ffde8-459a-4634-b567-5f900d40ceec-catalog-content\") pod \"certified-operators-jrfsv\" (UID: \"0c4ffde8-459a-4634-b567-5f900d40ceec\") " pod="openshift-marketplace/certified-operators-jrfsv" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.914311 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-879pq\" (UniqueName: \"kubernetes.io/projected/0c4ffde8-459a-4634-b567-5f900d40ceec-kube-api-access-879pq\") pod \"certified-operators-jrfsv\" (UID: \"0c4ffde8-459a-4634-b567-5f900d40ceec\") " pod="openshift-marketplace/certified-operators-jrfsv" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.915036 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c4ffde8-459a-4634-b567-5f900d40ceec-utilities\") pod \"certified-operators-jrfsv\" (UID: \"0c4ffde8-459a-4634-b567-5f900d40ceec\") " pod="openshift-marketplace/certified-operators-jrfsv" Dec 03 06:30:06 crc kubenswrapper[4818]: E1203 06:30:06.915302 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 06:30:07.415286937 +0000 UTC m=+165.106895689 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-zg4cb" (UID: "535884b5-264b-4471-962c-51579eeacab0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.915657 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c4ffde8-459a-4634-b567-5f900d40ceec-catalog-content\") pod \"certified-operators-jrfsv\" (UID: \"0c4ffde8-459a-4634-b567-5f900d40ceec\") " pod="openshift-marketplace/certified-operators-jrfsv" Dec 03 06:30:06 crc kubenswrapper[4818]: I1203 06:30:06.948441 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zfvqs"] Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.006207 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-879pq\" (UniqueName: \"kubernetes.io/projected/0c4ffde8-459a-4634-b567-5f900d40ceec-kube-api-access-879pq\") pod \"certified-operators-jrfsv\" (UID: \"0c4ffde8-459a-4634-b567-5f900d40ceec\") " pod="openshift-marketplace/certified-operators-jrfsv" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.018337 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.018530 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.018571 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-catalog-content\") pod \"community-operators-zfvqs\" (UID: \"46f85cb1-3840-4ff8-a506-d41af5a5f2f2\") " pod="openshift-marketplace/community-operators-zfvqs" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.018607 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2jx7\" (UniqueName: \"kubernetes.io/projected/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-kube-api-access-s2jx7\") pod \"community-operators-zfvqs\" (UID: \"46f85cb1-3840-4ff8-a506-d41af5a5f2f2\") " pod="openshift-marketplace/community-operators-zfvqs" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.018626 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-utilities\") pod \"community-operators-zfvqs\" (UID: \"46f85cb1-3840-4ff8-a506-d41af5a5f2f2\") " pod="openshift-marketplace/community-operators-zfvqs" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.018648 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:30:07 crc kubenswrapper[4818]: E1203 06:30:07.018748 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 06:30:07.518733346 +0000 UTC m=+165.210342098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.071482 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hrv77"] Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.072644 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hrv77" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.096412 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hrv77"] Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.111231 4818 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.111292 4818 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.120387 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.120450 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.120501 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-catalog-content\") pod \"community-operators-zfvqs\" (UID: \"46f85cb1-3840-4ff8-a506-d41af5a5f2f2\") " pod="openshift-marketplace/community-operators-zfvqs" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.120547 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2jx7\" (UniqueName: \"kubernetes.io/projected/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-kube-api-access-s2jx7\") pod \"community-operators-zfvqs\" (UID: \"46f85cb1-3840-4ff8-a506-d41af5a5f2f2\") " pod="openshift-marketplace/community-operators-zfvqs" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.120574 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-utilities\") pod \"community-operators-zfvqs\" (UID: \"46f85cb1-3840-4ff8-a506-d41af5a5f2f2\") " pod="openshift-marketplace/community-operators-zfvqs" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.120605 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.121272 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.121736 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-catalog-content\") pod \"community-operators-zfvqs\" (UID: \"46f85cb1-3840-4ff8-a506-d41af5a5f2f2\") " pod="openshift-marketplace/community-operators-zfvqs" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.122177 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-utilities\") pod \"community-operators-zfvqs\" (UID: \"46f85cb1-3840-4ff8-a506-d41af5a5f2f2\") " pod="openshift-marketplace/community-operators-zfvqs" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.145328 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.152901 4818 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.152937 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.153565 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2jx7\" (UniqueName: \"kubernetes.io/projected/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-kube-api-access-s2jx7\") pod \"community-operators-zfvqs\" (UID: \"46f85cb1-3840-4ff8-a506-d41af5a5f2f2\") " pod="openshift-marketplace/community-operators-zfvqs" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.191055 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.220155 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-zg4cb\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.221531 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc7nz\" (UniqueName: \"kubernetes.io/projected/366db0ca-c6f6-49ee-9968-05a60e74dae5-kube-api-access-qc7nz\") pod \"certified-operators-hrv77\" (UID: \"366db0ca-c6f6-49ee-9968-05a60e74dae5\") " pod="openshift-marketplace/certified-operators-hrv77" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.221611 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/366db0ca-c6f6-49ee-9968-05a60e74dae5-catalog-content\") pod \"certified-operators-hrv77\" (UID: \"366db0ca-c6f6-49ee-9968-05a60e74dae5\") " pod="openshift-marketplace/certified-operators-hrv77" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.221677 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/366db0ca-c6f6-49ee-9968-05a60e74dae5-utilities\") pod \"certified-operators-hrv77\" (UID: \"366db0ca-c6f6-49ee-9968-05a60e74dae5\") " pod="openshift-marketplace/certified-operators-hrv77" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.279649 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jrfsv" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.280297 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfvqs" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.284264 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.322621 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.323041 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/366db0ca-c6f6-49ee-9968-05a60e74dae5-catalog-content\") pod \"certified-operators-hrv77\" (UID: \"366db0ca-c6f6-49ee-9968-05a60e74dae5\") " pod="openshift-marketplace/certified-operators-hrv77" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.323139 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/366db0ca-c6f6-49ee-9968-05a60e74dae5-utilities\") pod \"certified-operators-hrv77\" (UID: \"366db0ca-c6f6-49ee-9968-05a60e74dae5\") " pod="openshift-marketplace/certified-operators-hrv77" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.323178 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc7nz\" (UniqueName: \"kubernetes.io/projected/366db0ca-c6f6-49ee-9968-05a60e74dae5-kube-api-access-qc7nz\") pod \"certified-operators-hrv77\" (UID: \"366db0ca-c6f6-49ee-9968-05a60e74dae5\") " pod="openshift-marketplace/certified-operators-hrv77" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.324235 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/366db0ca-c6f6-49ee-9968-05a60e74dae5-catalog-content\") pod \"certified-operators-hrv77\" (UID: \"366db0ca-c6f6-49ee-9968-05a60e74dae5\") " pod="openshift-marketplace/certified-operators-hrv77" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.324490 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/366db0ca-c6f6-49ee-9968-05a60e74dae5-utilities\") pod \"certified-operators-hrv77\" (UID: \"366db0ca-c6f6-49ee-9968-05a60e74dae5\") " pod="openshift-marketplace/certified-operators-hrv77" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.342287 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc7nz\" (UniqueName: \"kubernetes.io/projected/366db0ca-c6f6-49ee-9968-05a60e74dae5-kube-api-access-qc7nz\") pod \"certified-operators-hrv77\" (UID: \"366db0ca-c6f6-49ee-9968-05a60e74dae5\") " pod="openshift-marketplace/certified-operators-hrv77" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.396200 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hrv77" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.400157 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.551671 4818 generic.go:334] "Generic (PLEG): container finished" podID="f033be5e-5c71-4aeb-8825-bcf234d174a1" containerID="820336f7aa4564a8c9d5561a937f63298b3615bc679942f01a71a37b9a3328be" exitCode=0 Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.551762 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" event={"ID":"f033be5e-5c71-4aeb-8825-bcf234d174a1","Type":"ContainerDied","Data":"820336f7aa4564a8c9d5561a937f63298b3615bc679942f01a71a37b9a3328be"} Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.559286 4818 patch_prober.go:28] interesting pod/router-default-5444994796-fz577 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:30:07 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Dec 03 06:30:07 crc kubenswrapper[4818]: [+]process-running ok Dec 03 06:30:07 crc kubenswrapper[4818]: healthz check failed Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.559363 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fz577" podUID="05f7659c-85e9-4337-8741-9958f12895d4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.559539 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k7wwb"] Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.590223 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" event={"ID":"b14853e5-2388-4dba-a3ca-dd9453ea3452","Type":"ContainerStarted","Data":"88d75e94bd4a4dbf5bc0fe33be5ed033e4a0e1e6cd3b5a2ce248e822e7650ae2"} Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.744162 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-gkgqw" podStartSLOduration=11.743970632 podStartE2EDuration="11.743970632s" podCreationTimestamp="2025-12-03 06:29:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:07.628998426 +0000 UTC m=+165.320607178" watchObservedRunningTime="2025-12-03 06:30:07.743970632 +0000 UTC m=+165.435579384" Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.745577 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jrfsv"] Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.810642 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zfvqs"] Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.837905 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 06:30:07 crc kubenswrapper[4818]: W1203 06:30:07.843200 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46f85cb1_3840_4ff8_a506_d41af5a5f2f2.slice/crio-ae072f843b0ba8de6c4af5338aa15e157a03f03e1abe0155846faf6051e3c6f4 WatchSource:0}: Error finding container ae072f843b0ba8de6c4af5338aa15e157a03f03e1abe0155846faf6051e3c6f4: Status 404 returned error can't find the container with id ae072f843b0ba8de6c4af5338aa15e157a03f03e1abe0155846faf6051e3c6f4 Dec 03 06:30:07 crc kubenswrapper[4818]: I1203 06:30:07.919464 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zg4cb"] Dec 03 06:30:07 crc kubenswrapper[4818]: W1203 06:30:07.981135 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod535884b5_264b_4471_962c_51579eeacab0.slice/crio-0335d8d890192bf536f57525605b2579f9e0e5b041909ae284dc7bd718feffe1 WatchSource:0}: Error finding container 0335d8d890192bf536f57525605b2579f9e0e5b041909ae284dc7bd718feffe1: Status 404 returned error can't find the container with id 0335d8d890192bf536f57525605b2579f9e0e5b041909ae284dc7bd718feffe1 Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.017555 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hrv77"] Dec 03 06:30:08 crc kubenswrapper[4818]: W1203 06:30:08.028035 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod366db0ca_c6f6_49ee_9968_05a60e74dae5.slice/crio-aaed7d315eefa0f5c38cf0687aee9fa92846f337bf0557ddc7adfb90604dcdb0 WatchSource:0}: Error finding container aaed7d315eefa0f5c38cf0687aee9fa92846f337bf0557ddc7adfb90604dcdb0: Status 404 returned error can't find the container with id aaed7d315eefa0f5c38cf0687aee9fa92846f337bf0557ddc7adfb90604dcdb0 Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.344905 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs\") pod \"network-metrics-daemon-gq8r2\" (UID: \"2d8af736-67f3-4db0-a7a5-a24af4222b86\") " pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.350669 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d8af736-67f3-4db0-a7a5-a24af4222b86-metrics-certs\") pod \"network-metrics-daemon-gq8r2\" (UID: \"2d8af736-67f3-4db0-a7a5-a24af4222b86\") " pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.360889 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-gq8r2" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.510485 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-965hr"] Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.512905 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-965hr" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.515043 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.520868 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-965hr"] Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.555578 4818 patch_prober.go:28] interesting pod/router-default-5444994796-fz577 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:30:08 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Dec 03 06:30:08 crc kubenswrapper[4818]: [+]process-running ok Dec 03 06:30:08 crc kubenswrapper[4818]: healthz check failed Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.555631 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fz577" podUID="05f7659c-85e9-4337-8741-9958f12895d4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.595530 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91223086-535a-47c4-a843-3bb100a7da50-catalog-content\") pod \"redhat-marketplace-965hr\" (UID: \"91223086-535a-47c4-a843-3bb100a7da50\") " pod="openshift-marketplace/redhat-marketplace-965hr" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.595651 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91223086-535a-47c4-a843-3bb100a7da50-utilities\") pod \"redhat-marketplace-965hr\" (UID: \"91223086-535a-47c4-a843-3bb100a7da50\") " pod="openshift-marketplace/redhat-marketplace-965hr" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.595671 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw8b8\" (UniqueName: \"kubernetes.io/projected/91223086-535a-47c4-a843-3bb100a7da50-kube-api-access-tw8b8\") pod \"redhat-marketplace-965hr\" (UID: \"91223086-535a-47c4-a843-3bb100a7da50\") " pod="openshift-marketplace/redhat-marketplace-965hr" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.602898 4818 generic.go:334] "Generic (PLEG): container finished" podID="0c4ffde8-459a-4634-b567-5f900d40ceec" containerID="6d01faa4d0822c30f72b3f4b18f389157c34e498fefe338180d930d5af9bbfd2" exitCode=0 Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.602978 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrfsv" event={"ID":"0c4ffde8-459a-4634-b567-5f900d40ceec","Type":"ContainerDied","Data":"6d01faa4d0822c30f72b3f4b18f389157c34e498fefe338180d930d5af9bbfd2"} Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.603004 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrfsv" event={"ID":"0c4ffde8-459a-4634-b567-5f900d40ceec","Type":"ContainerStarted","Data":"a36a9c146ab72adba1f3bd6ba5135920ec58ce50478c9a9c11738cffafdcc878"} Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.604414 4818 generic.go:334] "Generic (PLEG): container finished" podID="46f85cb1-3840-4ff8-a506-d41af5a5f2f2" containerID="4b91fc3922ad91740eaaaed9e1c9bdb4bee8d9ae7028f086be79ee1f75691376" exitCode=0 Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.604935 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.605210 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfvqs" event={"ID":"46f85cb1-3840-4ff8-a506-d41af5a5f2f2","Type":"ContainerDied","Data":"4b91fc3922ad91740eaaaed9e1c9bdb4bee8d9ae7028f086be79ee1f75691376"} Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.605238 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfvqs" event={"ID":"46f85cb1-3840-4ff8-a506-d41af5a5f2f2","Type":"ContainerStarted","Data":"ae072f843b0ba8de6c4af5338aa15e157a03f03e1abe0155846faf6051e3c6f4"} Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.610304 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" event={"ID":"535884b5-264b-4471-962c-51579eeacab0","Type":"ContainerStarted","Data":"78bb80e7ab7e9c0fdbe26cb87cb807c0970253b99e1e0d90089eea4780144efb"} Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.610350 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" event={"ID":"535884b5-264b-4471-962c-51579eeacab0","Type":"ContainerStarted","Data":"0335d8d890192bf536f57525605b2579f9e0e5b041909ae284dc7bd718feffe1"} Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.612038 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.615037 4818 generic.go:334] "Generic (PLEG): container finished" podID="0301f892-ca1d-4f8b-a429-6d0ad96a50dc" containerID="28c1098bef6a1e692c39f2a2050ee72cb8efdb87e3e4987f112d3233edf99904" exitCode=0 Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.615103 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7wwb" event={"ID":"0301f892-ca1d-4f8b-a429-6d0ad96a50dc","Type":"ContainerDied","Data":"28c1098bef6a1e692c39f2a2050ee72cb8efdb87e3e4987f112d3233edf99904"} Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.615133 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7wwb" event={"ID":"0301f892-ca1d-4f8b-a429-6d0ad96a50dc","Type":"ContainerStarted","Data":"4625e08fc9cd6569bd4b04774a6c74dd03d178c2764532a534a5c5df30f3162c"} Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.619959 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-gq8r2"] Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.626017 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrv77" event={"ID":"366db0ca-c6f6-49ee-9968-05a60e74dae5","Type":"ContainerStarted","Data":"aaed7d315eefa0f5c38cf0687aee9fa92846f337bf0557ddc7adfb90604dcdb0"} Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.632844 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539","Type":"ContainerStarted","Data":"e4f63a3c35ec8e23560e177ab1423ef85aef4fb8fc40e4517a30c0cb1f3fc977"} Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.632930 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539","Type":"ContainerStarted","Data":"c6f95f2c6bc8a748463741643b5cee3f6c3ba108a092689e2440fd8d6ff957e0"} Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.668227 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" podStartSLOduration=142.668210572 podStartE2EDuration="2m22.668210572s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:08.666055008 +0000 UTC m=+166.357663760" watchObservedRunningTime="2025-12-03 06:30:08.668210572 +0000 UTC m=+166.359819324" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.696716 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91223086-535a-47c4-a843-3bb100a7da50-utilities\") pod \"redhat-marketplace-965hr\" (UID: \"91223086-535a-47c4-a843-3bb100a7da50\") " pod="openshift-marketplace/redhat-marketplace-965hr" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.696752 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw8b8\" (UniqueName: \"kubernetes.io/projected/91223086-535a-47c4-a843-3bb100a7da50-kube-api-access-tw8b8\") pod \"redhat-marketplace-965hr\" (UID: \"91223086-535a-47c4-a843-3bb100a7da50\") " pod="openshift-marketplace/redhat-marketplace-965hr" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.696785 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91223086-535a-47c4-a843-3bb100a7da50-catalog-content\") pod \"redhat-marketplace-965hr\" (UID: \"91223086-535a-47c4-a843-3bb100a7da50\") " pod="openshift-marketplace/redhat-marketplace-965hr" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.697244 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91223086-535a-47c4-a843-3bb100a7da50-catalog-content\") pod \"redhat-marketplace-965hr\" (UID: \"91223086-535a-47c4-a843-3bb100a7da50\") " pod="openshift-marketplace/redhat-marketplace-965hr" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.698071 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91223086-535a-47c4-a843-3bb100a7da50-utilities\") pod \"redhat-marketplace-965hr\" (UID: \"91223086-535a-47c4-a843-3bb100a7da50\") " pod="openshift-marketplace/redhat-marketplace-965hr" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.717266 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw8b8\" (UniqueName: \"kubernetes.io/projected/91223086-535a-47c4-a843-3bb100a7da50-kube-api-access-tw8b8\") pod \"redhat-marketplace-965hr\" (UID: \"91223086-535a-47c4-a843-3bb100a7da50\") " pod="openshift-marketplace/redhat-marketplace-965hr" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.779215 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.829290 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-965hr" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.835398 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.856295 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.856273324 podStartE2EDuration="2.856273324s" podCreationTimestamp="2025-12-03 06:30:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:08.698614607 +0000 UTC m=+166.390223359" watchObservedRunningTime="2025-12-03 06:30:08.856273324 +0000 UTC m=+166.547882076" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.858544 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xfzk6"] Dec 03 06:30:08 crc kubenswrapper[4818]: E1203 06:30:08.858772 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f033be5e-5c71-4aeb-8825-bcf234d174a1" containerName="collect-profiles" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.858789 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f033be5e-5c71-4aeb-8825-bcf234d174a1" containerName="collect-profiles" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.858959 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f033be5e-5c71-4aeb-8825-bcf234d174a1" containerName="collect-profiles" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.859808 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfzk6" Dec 03 06:30:08 crc kubenswrapper[4818]: I1203 06:30:08.868760 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfzk6"] Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.001385 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfmch\" (UniqueName: \"kubernetes.io/projected/f033be5e-5c71-4aeb-8825-bcf234d174a1-kube-api-access-lfmch\") pod \"f033be5e-5c71-4aeb-8825-bcf234d174a1\" (UID: \"f033be5e-5c71-4aeb-8825-bcf234d174a1\") " Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.001512 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f033be5e-5c71-4aeb-8825-bcf234d174a1-secret-volume\") pod \"f033be5e-5c71-4aeb-8825-bcf234d174a1\" (UID: \"f033be5e-5c71-4aeb-8825-bcf234d174a1\") " Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.001602 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f033be5e-5c71-4aeb-8825-bcf234d174a1-config-volume\") pod \"f033be5e-5c71-4aeb-8825-bcf234d174a1\" (UID: \"f033be5e-5c71-4aeb-8825-bcf234d174a1\") " Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.001856 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-catalog-content\") pod \"redhat-marketplace-xfzk6\" (UID: \"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d\") " pod="openshift-marketplace/redhat-marketplace-xfzk6" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.001905 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-utilities\") pod \"redhat-marketplace-xfzk6\" (UID: \"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d\") " pod="openshift-marketplace/redhat-marketplace-xfzk6" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.001937 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdt47\" (UniqueName: \"kubernetes.io/projected/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-kube-api-access-bdt47\") pod \"redhat-marketplace-xfzk6\" (UID: \"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d\") " pod="openshift-marketplace/redhat-marketplace-xfzk6" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.002426 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f033be5e-5c71-4aeb-8825-bcf234d174a1-config-volume" (OuterVolumeSpecName: "config-volume") pod "f033be5e-5c71-4aeb-8825-bcf234d174a1" (UID: "f033be5e-5c71-4aeb-8825-bcf234d174a1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.007641 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f033be5e-5c71-4aeb-8825-bcf234d174a1-kube-api-access-lfmch" (OuterVolumeSpecName: "kube-api-access-lfmch") pod "f033be5e-5c71-4aeb-8825-bcf234d174a1" (UID: "f033be5e-5c71-4aeb-8825-bcf234d174a1"). InnerVolumeSpecName "kube-api-access-lfmch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.007807 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f033be5e-5c71-4aeb-8825-bcf234d174a1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f033be5e-5c71-4aeb-8825-bcf234d174a1" (UID: "f033be5e-5c71-4aeb-8825-bcf234d174a1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.024461 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-965hr"] Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.103757 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-catalog-content\") pod \"redhat-marketplace-xfzk6\" (UID: \"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d\") " pod="openshift-marketplace/redhat-marketplace-xfzk6" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.103841 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-utilities\") pod \"redhat-marketplace-xfzk6\" (UID: \"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d\") " pod="openshift-marketplace/redhat-marketplace-xfzk6" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.103871 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdt47\" (UniqueName: \"kubernetes.io/projected/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-kube-api-access-bdt47\") pod \"redhat-marketplace-xfzk6\" (UID: \"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d\") " pod="openshift-marketplace/redhat-marketplace-xfzk6" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.103921 4818 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f033be5e-5c71-4aeb-8825-bcf234d174a1-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.103939 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfmch\" (UniqueName: \"kubernetes.io/projected/f033be5e-5c71-4aeb-8825-bcf234d174a1-kube-api-access-lfmch\") on node \"crc\" DevicePath \"\"" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.103952 4818 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f033be5e-5c71-4aeb-8825-bcf234d174a1-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.104425 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-utilities\") pod \"redhat-marketplace-xfzk6\" (UID: \"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d\") " pod="openshift-marketplace/redhat-marketplace-xfzk6" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.105843 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-catalog-content\") pod \"redhat-marketplace-xfzk6\" (UID: \"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d\") " pod="openshift-marketplace/redhat-marketplace-xfzk6" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.121698 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdt47\" (UniqueName: \"kubernetes.io/projected/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-kube-api-access-bdt47\") pod \"redhat-marketplace-xfzk6\" (UID: \"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d\") " pod="openshift-marketplace/redhat-marketplace-xfzk6" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.275137 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfzk6" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.454760 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfzk6"] Dec 03 06:30:09 crc kubenswrapper[4818]: W1203 06:30:09.465044 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae7bd5bd_4cac_44e9_a4c6_da6de50a1c8d.slice/crio-ba57498bff52ce47e338363a843becdd6295fcf2ec3ca636dbb8fe453930e5d2 WatchSource:0}: Error finding container ba57498bff52ce47e338363a843becdd6295fcf2ec3ca636dbb8fe453930e5d2: Status 404 returned error can't find the container with id ba57498bff52ce47e338363a843becdd6295fcf2ec3ca636dbb8fe453930e5d2 Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.552805 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.555442 4818 patch_prober.go:28] interesting pod/router-default-5444994796-fz577 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:30:09 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Dec 03 06:30:09 crc kubenswrapper[4818]: [+]process-running ok Dec 03 06:30:09 crc kubenswrapper[4818]: healthz check failed Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.555498 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fz577" podUID="05f7659c-85e9-4337-8741-9958f12895d4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.642513 4818 generic.go:334] "Generic (PLEG): container finished" podID="366db0ca-c6f6-49ee-9968-05a60e74dae5" containerID="7caa155be3fa4a7bc0fa4329fcb2e2f4b74412181d26ee1ba285e65a7d136de5" exitCode=0 Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.642750 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrv77" event={"ID":"366db0ca-c6f6-49ee-9968-05a60e74dae5","Type":"ContainerDied","Data":"7caa155be3fa4a7bc0fa4329fcb2e2f4b74412181d26ee1ba285e65a7d136de5"} Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.645209 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" event={"ID":"f033be5e-5c71-4aeb-8825-bcf234d174a1","Type":"ContainerDied","Data":"dbd2e588851182739b2867e2911ed72b1374191776428f862721d1bc397c4dde"} Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.645241 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbd2e588851182739b2867e2911ed72b1374191776428f862721d1bc397c4dde" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.645300 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.648312 4818 generic.go:334] "Generic (PLEG): container finished" podID="e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539" containerID="e4f63a3c35ec8e23560e177ab1423ef85aef4fb8fc40e4517a30c0cb1f3fc977" exitCode=0 Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.648382 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539","Type":"ContainerDied","Data":"e4f63a3c35ec8e23560e177ab1423ef85aef4fb8fc40e4517a30c0cb1f3fc977"} Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.651630 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfzk6" event={"ID":"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d","Type":"ContainerStarted","Data":"8345a9f57c7efc24bd7ff9af58ee1afb3e2fc140acab644908909991f7f9fc8a"} Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.651673 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfzk6" event={"ID":"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d","Type":"ContainerStarted","Data":"ba57498bff52ce47e338363a843becdd6295fcf2ec3ca636dbb8fe453930e5d2"} Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.656058 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.656748 4818 generic.go:334] "Generic (PLEG): container finished" podID="91223086-535a-47c4-a843-3bb100a7da50" containerID="d53dac0a6d42f7ce63071a532981951505c9038d78d08c70cfe80d0d6bf73037" exitCode=0 Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.656958 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-965hr" event={"ID":"91223086-535a-47c4-a843-3bb100a7da50","Type":"ContainerDied","Data":"d53dac0a6d42f7ce63071a532981951505c9038d78d08c70cfe80d0d6bf73037"} Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.657019 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.657031 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-965hr" event={"ID":"91223086-535a-47c4-a843-3bb100a7da50","Type":"ContainerStarted","Data":"3c57e09323739e67935be77ba6d4e194f10e027bfad4f009931b9eca389a8aa0"} Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.657782 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x7vlw"] Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.658610 4818 patch_prober.go:28] interesting pod/console-f9d7485db-sccrw container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.658661 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-sccrw" podUID="6d2839d4-e624-4d61-9225-894a998e9e9e" containerName="console" probeResult="failure" output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.659014 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7vlw" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.661045 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.666303 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" event={"ID":"2d8af736-67f3-4db0-a7a5-a24af4222b86","Type":"ContainerStarted","Data":"dd8a436774ec6719d3b11f7c2beffa14fa8865ef4824661bdec908302a708bb5"} Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.666338 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" event={"ID":"2d8af736-67f3-4db0-a7a5-a24af4222b86","Type":"ContainerStarted","Data":"83acdae74a3cdfb040e8a56a3f8ce32783cdcd9be258d1ada6e3829902aa45a5"} Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.666350 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-gq8r2" event={"ID":"2d8af736-67f3-4db0-a7a5-a24af4222b86","Type":"ContainerStarted","Data":"787e4375722081bc4e0b5455eea477bc0a989bd8f7ffc12db31281d47ab4e725"} Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.671195 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7vlw"] Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.696444 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.696519 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.708115 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.727480 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p2fj\" (UniqueName: \"kubernetes.io/projected/75955bfe-ae6e-481d-84b5-3de486a86827-kube-api-access-6p2fj\") pod \"redhat-operators-x7vlw\" (UID: \"75955bfe-ae6e-481d-84b5-3de486a86827\") " pod="openshift-marketplace/redhat-operators-x7vlw" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.727527 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75955bfe-ae6e-481d-84b5-3de486a86827-catalog-content\") pod \"redhat-operators-x7vlw\" (UID: \"75955bfe-ae6e-481d-84b5-3de486a86827\") " pod="openshift-marketplace/redhat-operators-x7vlw" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.727596 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75955bfe-ae6e-481d-84b5-3de486a86827-utilities\") pod \"redhat-operators-x7vlw\" (UID: \"75955bfe-ae6e-481d-84b5-3de486a86827\") " pod="openshift-marketplace/redhat-operators-x7vlw" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.762092 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-gq8r2" podStartSLOduration=143.762073786 podStartE2EDuration="2m23.762073786s" podCreationTimestamp="2025-12-03 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:09.760349703 +0000 UTC m=+167.451958455" watchObservedRunningTime="2025-12-03 06:30:09.762073786 +0000 UTC m=+167.453682538" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.776873 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-pvb2q container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.776927 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-pvb2q" podUID="4b90f9f4-fcfa-46b3-80de-b8e335c84e5b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.776947 4818 patch_prober.go:28] interesting pod/downloads-7954f5f757-pvb2q container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.777061 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-pvb2q" podUID="4b90f9f4-fcfa-46b3-80de-b8e335c84e5b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.830969 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p2fj\" (UniqueName: \"kubernetes.io/projected/75955bfe-ae6e-481d-84b5-3de486a86827-kube-api-access-6p2fj\") pod \"redhat-operators-x7vlw\" (UID: \"75955bfe-ae6e-481d-84b5-3de486a86827\") " pod="openshift-marketplace/redhat-operators-x7vlw" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.831014 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75955bfe-ae6e-481d-84b5-3de486a86827-catalog-content\") pod \"redhat-operators-x7vlw\" (UID: \"75955bfe-ae6e-481d-84b5-3de486a86827\") " pod="openshift-marketplace/redhat-operators-x7vlw" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.831123 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75955bfe-ae6e-481d-84b5-3de486a86827-utilities\") pod \"redhat-operators-x7vlw\" (UID: \"75955bfe-ae6e-481d-84b5-3de486a86827\") " pod="openshift-marketplace/redhat-operators-x7vlw" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.833337 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75955bfe-ae6e-481d-84b5-3de486a86827-catalog-content\") pod \"redhat-operators-x7vlw\" (UID: \"75955bfe-ae6e-481d-84b5-3de486a86827\") " pod="openshift-marketplace/redhat-operators-x7vlw" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.833450 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75955bfe-ae6e-481d-84b5-3de486a86827-utilities\") pod \"redhat-operators-x7vlw\" (UID: \"75955bfe-ae6e-481d-84b5-3de486a86827\") " pod="openshift-marketplace/redhat-operators-x7vlw" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.857044 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p2fj\" (UniqueName: \"kubernetes.io/projected/75955bfe-ae6e-481d-84b5-3de486a86827-kube-api-access-6p2fj\") pod \"redhat-operators-x7vlw\" (UID: \"75955bfe-ae6e-481d-84b5-3de486a86827\") " pod="openshift-marketplace/redhat-operators-x7vlw" Dec 03 06:30:09 crc kubenswrapper[4818]: I1203 06:30:09.917419 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.028252 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7vlw" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.060773 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fzf6r"] Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.061790 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fzf6r" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.090521 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fzf6r"] Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.237432 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b81e21fd-b361-4853-9a91-e034667fce63-utilities\") pod \"redhat-operators-fzf6r\" (UID: \"b81e21fd-b361-4853-9a91-e034667fce63\") " pod="openshift-marketplace/redhat-operators-fzf6r" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.237478 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpgzf\" (UniqueName: \"kubernetes.io/projected/b81e21fd-b361-4853-9a91-e034667fce63-kube-api-access-vpgzf\") pod \"redhat-operators-fzf6r\" (UID: \"b81e21fd-b361-4853-9a91-e034667fce63\") " pod="openshift-marketplace/redhat-operators-fzf6r" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.237553 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b81e21fd-b361-4853-9a91-e034667fce63-catalog-content\") pod \"redhat-operators-fzf6r\" (UID: \"b81e21fd-b361-4853-9a91-e034667fce63\") " pod="openshift-marketplace/redhat-operators-fzf6r" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.336269 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7vlw"] Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.340046 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b81e21fd-b361-4853-9a91-e034667fce63-utilities\") pod \"redhat-operators-fzf6r\" (UID: \"b81e21fd-b361-4853-9a91-e034667fce63\") " pod="openshift-marketplace/redhat-operators-fzf6r" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.340092 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpgzf\" (UniqueName: \"kubernetes.io/projected/b81e21fd-b361-4853-9a91-e034667fce63-kube-api-access-vpgzf\") pod \"redhat-operators-fzf6r\" (UID: \"b81e21fd-b361-4853-9a91-e034667fce63\") " pod="openshift-marketplace/redhat-operators-fzf6r" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.340153 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b81e21fd-b361-4853-9a91-e034667fce63-catalog-content\") pod \"redhat-operators-fzf6r\" (UID: \"b81e21fd-b361-4853-9a91-e034667fce63\") " pod="openshift-marketplace/redhat-operators-fzf6r" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.340695 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b81e21fd-b361-4853-9a91-e034667fce63-catalog-content\") pod \"redhat-operators-fzf6r\" (UID: \"b81e21fd-b361-4853-9a91-e034667fce63\") " pod="openshift-marketplace/redhat-operators-fzf6r" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.340937 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b81e21fd-b361-4853-9a91-e034667fce63-utilities\") pod \"redhat-operators-fzf6r\" (UID: \"b81e21fd-b361-4853-9a91-e034667fce63\") " pod="openshift-marketplace/redhat-operators-fzf6r" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.361557 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpgzf\" (UniqueName: \"kubernetes.io/projected/b81e21fd-b361-4853-9a91-e034667fce63-kube-api-access-vpgzf\") pod \"redhat-operators-fzf6r\" (UID: \"b81e21fd-b361-4853-9a91-e034667fce63\") " pod="openshift-marketplace/redhat-operators-fzf6r" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.431154 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fzf6r" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.571281 4818 patch_prober.go:28] interesting pod/router-default-5444994796-fz577 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:30:10 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Dec 03 06:30:10 crc kubenswrapper[4818]: [+]process-running ok Dec 03 06:30:10 crc kubenswrapper[4818]: healthz check failed Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.571323 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fz577" podUID="05f7659c-85e9-4337-8741-9958f12895d4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.675711 4818 generic.go:334] "Generic (PLEG): container finished" podID="75955bfe-ae6e-481d-84b5-3de486a86827" containerID="4736d6686b2ae0fdb52c2f22ea23a0f6305b23829fb928b14edd686c66950d05" exitCode=0 Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.675808 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7vlw" event={"ID":"75955bfe-ae6e-481d-84b5-3de486a86827","Type":"ContainerDied","Data":"4736d6686b2ae0fdb52c2f22ea23a0f6305b23829fb928b14edd686c66950d05"} Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.676162 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7vlw" event={"ID":"75955bfe-ae6e-481d-84b5-3de486a86827","Type":"ContainerStarted","Data":"fb107d86eb8bccc5f47d4ffdc35eaf554477a2a0592d8ae5d3629ab98e18e1f4"} Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.685659 4818 generic.go:334] "Generic (PLEG): container finished" podID="ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d" containerID="8345a9f57c7efc24bd7ff9af58ee1afb3e2fc140acab644908909991f7f9fc8a" exitCode=0 Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.686991 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfzk6" event={"ID":"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d","Type":"ContainerDied","Data":"8345a9f57c7efc24bd7ff9af58ee1afb3e2fc140acab644908909991f7f9fc8a"} Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.715631 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-qnxnn" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.903336 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.904323 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.921839 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.922032 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.924404 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.985480 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ab18c1d7-b656-4949-96a3-c283b6b1ebfa-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ab18c1d7-b656-4949-96a3-c283b6b1ebfa\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:30:10 crc kubenswrapper[4818]: I1203 06:30:10.985544 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ab18c1d7-b656-4949-96a3-c283b6b1ebfa-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ab18c1d7-b656-4949-96a3-c283b6b1ebfa\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.023266 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fzf6r"] Dec 03 06:30:11 crc kubenswrapper[4818]: W1203 06:30:11.084941 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb81e21fd_b361_4853_9a91_e034667fce63.slice/crio-217dfb9b5dc13e3b154236214528fe2fb8085b020b3110f7a2e5380419bd1e9d WatchSource:0}: Error finding container 217dfb9b5dc13e3b154236214528fe2fb8085b020b3110f7a2e5380419bd1e9d: Status 404 returned error can't find the container with id 217dfb9b5dc13e3b154236214528fe2fb8085b020b3110f7a2e5380419bd1e9d Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.087567 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ab18c1d7-b656-4949-96a3-c283b6b1ebfa-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ab18c1d7-b656-4949-96a3-c283b6b1ebfa\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.087624 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ab18c1d7-b656-4949-96a3-c283b6b1ebfa-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ab18c1d7-b656-4949-96a3-c283b6b1ebfa\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.088069 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ab18c1d7-b656-4949-96a3-c283b6b1ebfa-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ab18c1d7-b656-4949-96a3-c283b6b1ebfa\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.123543 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ab18c1d7-b656-4949-96a3-c283b6b1ebfa-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ab18c1d7-b656-4949-96a3-c283b6b1ebfa\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.263297 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.361161 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.503286 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539-kubelet-dir\") pod \"e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539\" (UID: \"e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539\") " Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.503349 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539-kube-api-access\") pod \"e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539\" (UID: \"e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539\") " Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.503419 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539" (UID: "e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.503808 4818 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.565385 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539" (UID: "e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.565927 4818 patch_prober.go:28] interesting pod/router-default-5444994796-fz577 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:30:11 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Dec 03 06:30:11 crc kubenswrapper[4818]: [+]process-running ok Dec 03 06:30:11 crc kubenswrapper[4818]: healthz check failed Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.565967 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fz577" podUID="05f7659c-85e9-4337-8741-9958f12895d4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.604682 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.695495 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 06:30:11 crc kubenswrapper[4818]: W1203 06:30:11.723331 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podab18c1d7_b656_4949_96a3_c283b6b1ebfa.slice/crio-cf2f71863f4b354c5782076fe77bc46b8aa501fa403b7284d29c4890ef51bc9a WatchSource:0}: Error finding container cf2f71863f4b354c5782076fe77bc46b8aa501fa403b7284d29c4890ef51bc9a: Status 404 returned error can't find the container with id cf2f71863f4b354c5782076fe77bc46b8aa501fa403b7284d29c4890ef51bc9a Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.727249 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.727252 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539","Type":"ContainerDied","Data":"c6f95f2c6bc8a748463741643b5cee3f6c3ba108a092689e2440fd8d6ff957e0"} Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.727387 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6f95f2c6bc8a748463741643b5cee3f6c3ba108a092689e2440fd8d6ff957e0" Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.735453 4818 generic.go:334] "Generic (PLEG): container finished" podID="b81e21fd-b361-4853-9a91-e034667fce63" containerID="879661c34ad5ece5a71069376c8adc67fea3f6dbc8d6e5dc4559259ee15a180d" exitCode=0 Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.735572 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fzf6r" event={"ID":"b81e21fd-b361-4853-9a91-e034667fce63","Type":"ContainerDied","Data":"879661c34ad5ece5a71069376c8adc67fea3f6dbc8d6e5dc4559259ee15a180d"} Dec 03 06:30:11 crc kubenswrapper[4818]: I1203 06:30:11.735633 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fzf6r" event={"ID":"b81e21fd-b361-4853-9a91-e034667fce63","Type":"ContainerStarted","Data":"217dfb9b5dc13e3b154236214528fe2fb8085b020b3110f7a2e5380419bd1e9d"} Dec 03 06:30:12 crc kubenswrapper[4818]: I1203 06:30:12.555265 4818 patch_prober.go:28] interesting pod/router-default-5444994796-fz577 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:30:12 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Dec 03 06:30:12 crc kubenswrapper[4818]: [+]process-running ok Dec 03 06:30:12 crc kubenswrapper[4818]: healthz check failed Dec 03 06:30:12 crc kubenswrapper[4818]: I1203 06:30:12.555349 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fz577" podUID="05f7659c-85e9-4337-8741-9958f12895d4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:30:12 crc kubenswrapper[4818]: I1203 06:30:12.774491 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ab18c1d7-b656-4949-96a3-c283b6b1ebfa","Type":"ContainerStarted","Data":"b70ce109f23f6ac8f2e26e07112767e434d9963c0b79187c5c903f6e6b6dda71"} Dec 03 06:30:12 crc kubenswrapper[4818]: I1203 06:30:12.774921 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ab18c1d7-b656-4949-96a3-c283b6b1ebfa","Type":"ContainerStarted","Data":"cf2f71863f4b354c5782076fe77bc46b8aa501fa403b7284d29c4890ef51bc9a"} Dec 03 06:30:13 crc kubenswrapper[4818]: I1203 06:30:13.302346 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:30:13 crc kubenswrapper[4818]: I1203 06:30:13.302407 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:30:13 crc kubenswrapper[4818]: I1203 06:30:13.556719 4818 patch_prober.go:28] interesting pod/router-default-5444994796-fz577 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:30:13 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Dec 03 06:30:13 crc kubenswrapper[4818]: [+]process-running ok Dec 03 06:30:13 crc kubenswrapper[4818]: healthz check failed Dec 03 06:30:13 crc kubenswrapper[4818]: I1203 06:30:13.556770 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fz577" podUID="05f7659c-85e9-4337-8741-9958f12895d4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:30:13 crc kubenswrapper[4818]: I1203 06:30:13.792606 4818 generic.go:334] "Generic (PLEG): container finished" podID="ab18c1d7-b656-4949-96a3-c283b6b1ebfa" containerID="b70ce109f23f6ac8f2e26e07112767e434d9963c0b79187c5c903f6e6b6dda71" exitCode=0 Dec 03 06:30:13 crc kubenswrapper[4818]: I1203 06:30:13.792682 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ab18c1d7-b656-4949-96a3-c283b6b1ebfa","Type":"ContainerDied","Data":"b70ce109f23f6ac8f2e26e07112767e434d9963c0b79187c5c903f6e6b6dda71"} Dec 03 06:30:14 crc kubenswrapper[4818]: I1203 06:30:14.642741 4818 patch_prober.go:28] interesting pod/router-default-5444994796-fz577 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:30:14 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Dec 03 06:30:14 crc kubenswrapper[4818]: [+]process-running ok Dec 03 06:30:14 crc kubenswrapper[4818]: healthz check failed Dec 03 06:30:14 crc kubenswrapper[4818]: I1203 06:30:14.642797 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fz577" podUID="05f7659c-85e9-4337-8741-9958f12895d4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:30:15 crc kubenswrapper[4818]: I1203 06:30:15.007145 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-79bwt" Dec 03 06:30:15 crc kubenswrapper[4818]: I1203 06:30:15.246927 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:30:15 crc kubenswrapper[4818]: I1203 06:30:15.443999 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ab18c1d7-b656-4949-96a3-c283b6b1ebfa-kubelet-dir\") pod \"ab18c1d7-b656-4949-96a3-c283b6b1ebfa\" (UID: \"ab18c1d7-b656-4949-96a3-c283b6b1ebfa\") " Dec 03 06:30:15 crc kubenswrapper[4818]: I1203 06:30:15.444060 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ab18c1d7-b656-4949-96a3-c283b6b1ebfa-kube-api-access\") pod \"ab18c1d7-b656-4949-96a3-c283b6b1ebfa\" (UID: \"ab18c1d7-b656-4949-96a3-c283b6b1ebfa\") " Dec 03 06:30:15 crc kubenswrapper[4818]: I1203 06:30:15.444180 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab18c1d7-b656-4949-96a3-c283b6b1ebfa-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ab18c1d7-b656-4949-96a3-c283b6b1ebfa" (UID: "ab18c1d7-b656-4949-96a3-c283b6b1ebfa"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:30:15 crc kubenswrapper[4818]: I1203 06:30:15.444554 4818 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ab18c1d7-b656-4949-96a3-c283b6b1ebfa-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:30:15 crc kubenswrapper[4818]: I1203 06:30:15.450001 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab18c1d7-b656-4949-96a3-c283b6b1ebfa-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ab18c1d7-b656-4949-96a3-c283b6b1ebfa" (UID: "ab18c1d7-b656-4949-96a3-c283b6b1ebfa"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:30:15 crc kubenswrapper[4818]: I1203 06:30:15.545913 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ab18c1d7-b656-4949-96a3-c283b6b1ebfa-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 06:30:15 crc kubenswrapper[4818]: I1203 06:30:15.554171 4818 patch_prober.go:28] interesting pod/router-default-5444994796-fz577 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:30:15 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Dec 03 06:30:15 crc kubenswrapper[4818]: [+]process-running ok Dec 03 06:30:15 crc kubenswrapper[4818]: healthz check failed Dec 03 06:30:15 crc kubenswrapper[4818]: I1203 06:30:15.554262 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fz577" podUID="05f7659c-85e9-4337-8741-9958f12895d4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:30:15 crc kubenswrapper[4818]: I1203 06:30:15.808576 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ab18c1d7-b656-4949-96a3-c283b6b1ebfa","Type":"ContainerDied","Data":"cf2f71863f4b354c5782076fe77bc46b8aa501fa403b7284d29c4890ef51bc9a"} Dec 03 06:30:15 crc kubenswrapper[4818]: I1203 06:30:15.808623 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf2f71863f4b354c5782076fe77bc46b8aa501fa403b7284d29c4890ef51bc9a" Dec 03 06:30:15 crc kubenswrapper[4818]: I1203 06:30:15.808703 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 06:30:16 crc kubenswrapper[4818]: I1203 06:30:16.555098 4818 patch_prober.go:28] interesting pod/router-default-5444994796-fz577 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 06:30:16 crc kubenswrapper[4818]: [-]has-synced failed: reason withheld Dec 03 06:30:16 crc kubenswrapper[4818]: [+]process-running ok Dec 03 06:30:16 crc kubenswrapper[4818]: healthz check failed Dec 03 06:30:16 crc kubenswrapper[4818]: I1203 06:30:16.555167 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-fz577" podUID="05f7659c-85e9-4337-8741-9958f12895d4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:30:17 crc kubenswrapper[4818]: I1203 06:30:17.555981 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:30:17 crc kubenswrapper[4818]: I1203 06:30:17.561928 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-fz577" Dec 03 06:30:19 crc kubenswrapper[4818]: I1203 06:30:19.754088 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:30:19 crc kubenswrapper[4818]: I1203 06:30:19.758369 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:30:19 crc kubenswrapper[4818]: I1203 06:30:19.797755 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-pvb2q" Dec 03 06:30:27 crc kubenswrapper[4818]: I1203 06:30:27.297475 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:30:39 crc kubenswrapper[4818]: I1203 06:30:39.979609 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-6glnm" Dec 03 06:30:43 crc kubenswrapper[4818]: I1203 06:30:43.303032 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:30:43 crc kubenswrapper[4818]: I1203 06:30:43.303562 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:30:47 crc kubenswrapper[4818]: E1203 06:30:47.796570 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 06:30:47 crc kubenswrapper[4818]: E1203 06:30:47.797087 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vpgzf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-fzf6r_openshift-marketplace(b81e21fd-b361-4853-9a91-e034667fce63): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 06:30:47 crc kubenswrapper[4818]: E1203 06:30:47.798299 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-fzf6r" podUID="b81e21fd-b361-4853-9a91-e034667fce63" Dec 03 06:30:47 crc kubenswrapper[4818]: E1203 06:30:47.820020 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 06:30:47 crc kubenswrapper[4818]: E1203 06:30:47.820198 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6p2fj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-x7vlw_openshift-marketplace(75955bfe-ae6e-481d-84b5-3de486a86827): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 06:30:47 crc kubenswrapper[4818]: E1203 06:30:47.821448 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-x7vlw" podUID="75955bfe-ae6e-481d-84b5-3de486a86827" Dec 03 06:30:48 crc kubenswrapper[4818]: I1203 06:30:48.980626 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 06:30:48 crc kubenswrapper[4818]: E1203 06:30:48.981346 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab18c1d7-b656-4949-96a3-c283b6b1ebfa" containerName="pruner" Dec 03 06:30:48 crc kubenswrapper[4818]: I1203 06:30:48.981357 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab18c1d7-b656-4949-96a3-c283b6b1ebfa" containerName="pruner" Dec 03 06:30:48 crc kubenswrapper[4818]: E1203 06:30:48.981366 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539" containerName="pruner" Dec 03 06:30:48 crc kubenswrapper[4818]: I1203 06:30:48.981372 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539" containerName="pruner" Dec 03 06:30:48 crc kubenswrapper[4818]: I1203 06:30:48.981472 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab18c1d7-b656-4949-96a3-c283b6b1ebfa" containerName="pruner" Dec 03 06:30:48 crc kubenswrapper[4818]: I1203 06:30:48.981484 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0d0f0fd-4b4d-43bd-8138-8bf27e0ab539" containerName="pruner" Dec 03 06:30:48 crc kubenswrapper[4818]: I1203 06:30:48.981968 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:30:48 crc kubenswrapper[4818]: I1203 06:30:48.982126 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 06:30:48 crc kubenswrapper[4818]: I1203 06:30:48.985257 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 06:30:48 crc kubenswrapper[4818]: I1203 06:30:48.986073 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 06:30:49 crc kubenswrapper[4818]: E1203 06:30:49.082160 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-x7vlw" podUID="75955bfe-ae6e-481d-84b5-3de486a86827" Dec 03 06:30:49 crc kubenswrapper[4818]: E1203 06:30:49.082215 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-fzf6r" podUID="b81e21fd-b361-4853-9a91-e034667fce63" Dec 03 06:30:49 crc kubenswrapper[4818]: I1203 06:30:49.088331 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e845a73a-f851-421e-8708-55e618d3a11d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e845a73a-f851-421e-8708-55e618d3a11d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:30:49 crc kubenswrapper[4818]: I1203 06:30:49.088511 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e845a73a-f851-421e-8708-55e618d3a11d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e845a73a-f851-421e-8708-55e618d3a11d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:30:49 crc kubenswrapper[4818]: E1203 06:30:49.142236 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 06:30:49 crc kubenswrapper[4818]: E1203 06:30:49.142435 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-879pq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-jrfsv_openshift-marketplace(0c4ffde8-459a-4634-b567-5f900d40ceec): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 06:30:49 crc kubenswrapper[4818]: E1203 06:30:49.143644 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-jrfsv" podUID="0c4ffde8-459a-4634-b567-5f900d40ceec" Dec 03 06:30:49 crc kubenswrapper[4818]: I1203 06:30:49.189986 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e845a73a-f851-421e-8708-55e618d3a11d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e845a73a-f851-421e-8708-55e618d3a11d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:30:49 crc kubenswrapper[4818]: I1203 06:30:49.190053 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e845a73a-f851-421e-8708-55e618d3a11d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e845a73a-f851-421e-8708-55e618d3a11d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:30:49 crc kubenswrapper[4818]: I1203 06:30:49.190479 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e845a73a-f851-421e-8708-55e618d3a11d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e845a73a-f851-421e-8708-55e618d3a11d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:30:49 crc kubenswrapper[4818]: I1203 06:30:49.223193 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e845a73a-f851-421e-8708-55e618d3a11d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e845a73a-f851-421e-8708-55e618d3a11d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:30:49 crc kubenswrapper[4818]: I1203 06:30:49.308351 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:30:50 crc kubenswrapper[4818]: E1203 06:30:50.059315 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-jrfsv" podUID="0c4ffde8-459a-4634-b567-5f900d40ceec" Dec 03 06:30:50 crc kubenswrapper[4818]: E1203 06:30:50.119079 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 06:30:50 crc kubenswrapper[4818]: E1203 06:30:50.119285 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bdt47,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-xfzk6_openshift-marketplace(ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 06:30:50 crc kubenswrapper[4818]: E1203 06:30:50.120495 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-xfzk6" podUID="ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d" Dec 03 06:30:50 crc kubenswrapper[4818]: E1203 06:30:50.133680 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 06:30:50 crc kubenswrapper[4818]: E1203 06:30:50.133850 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tw8b8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-965hr_openshift-marketplace(91223086-535a-47c4-a843-3bb100a7da50): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 06:30:50 crc kubenswrapper[4818]: E1203 06:30:50.135019 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-965hr" podUID="91223086-535a-47c4-a843-3bb100a7da50" Dec 03 06:30:50 crc kubenswrapper[4818]: E1203 06:30:50.162313 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 06:30:50 crc kubenswrapper[4818]: E1203 06:30:50.162478 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qc7nz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-hrv77_openshift-marketplace(366db0ca-c6f6-49ee-9968-05a60e74dae5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 06:30:50 crc kubenswrapper[4818]: E1203 06:30:50.164424 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-hrv77" podUID="366db0ca-c6f6-49ee-9968-05a60e74dae5" Dec 03 06:30:51 crc kubenswrapper[4818]: E1203 06:30:51.382254 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-xfzk6" podUID="ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d" Dec 03 06:30:51 crc kubenswrapper[4818]: E1203 06:30:51.382723 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-hrv77" podUID="366db0ca-c6f6-49ee-9968-05a60e74dae5" Dec 03 06:30:51 crc kubenswrapper[4818]: E1203 06:30:51.382952 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-965hr" podUID="91223086-535a-47c4-a843-3bb100a7da50" Dec 03 06:30:51 crc kubenswrapper[4818]: E1203 06:30:51.451311 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 06:30:51 crc kubenswrapper[4818]: E1203 06:30:51.451551 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2jx7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zfvqs_openshift-marketplace(46f85cb1-3840-4ff8-a506-d41af5a5f2f2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 06:30:51 crc kubenswrapper[4818]: E1203 06:30:51.453195 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zfvqs" podUID="46f85cb1-3840-4ff8-a506-d41af5a5f2f2" Dec 03 06:30:51 crc kubenswrapper[4818]: E1203 06:30:51.475667 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 06:30:51 crc kubenswrapper[4818]: E1203 06:30:51.476151 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tscsg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-k7wwb_openshift-marketplace(0301f892-ca1d-4f8b-a429-6d0ad96a50dc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 06:30:51 crc kubenswrapper[4818]: E1203 06:30:51.477950 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-k7wwb" podUID="0301f892-ca1d-4f8b-a429-6d0ad96a50dc" Dec 03 06:30:51 crc kubenswrapper[4818]: I1203 06:30:51.801953 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 06:30:52 crc kubenswrapper[4818]: I1203 06:30:52.186881 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e845a73a-f851-421e-8708-55e618d3a11d","Type":"ContainerStarted","Data":"c7b11145e3f0872d35322f0ba788189a9e35ab9c55c0727acc2b12f6c6b5c3ba"} Dec 03 06:30:52 crc kubenswrapper[4818]: E1203 06:30:52.188983 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zfvqs" podUID="46f85cb1-3840-4ff8-a506-d41af5a5f2f2" Dec 03 06:30:52 crc kubenswrapper[4818]: E1203 06:30:52.189036 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-k7wwb" podUID="0301f892-ca1d-4f8b-a429-6d0ad96a50dc" Dec 03 06:30:53 crc kubenswrapper[4818]: I1203 06:30:53.193864 4818 generic.go:334] "Generic (PLEG): container finished" podID="e845a73a-f851-421e-8708-55e618d3a11d" containerID="5c8576251637b1a6f43c2e4961ce94a6e3282efd03fa78d3338d3ca53e3f6988" exitCode=0 Dec 03 06:30:53 crc kubenswrapper[4818]: I1203 06:30:53.193904 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e845a73a-f851-421e-8708-55e618d3a11d","Type":"ContainerDied","Data":"5c8576251637b1a6f43c2e4961ce94a6e3282efd03fa78d3338d3ca53e3f6988"} Dec 03 06:30:54 crc kubenswrapper[4818]: I1203 06:30:54.415437 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:30:54 crc kubenswrapper[4818]: I1203 06:30:54.456124 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e845a73a-f851-421e-8708-55e618d3a11d-kubelet-dir\") pod \"e845a73a-f851-421e-8708-55e618d3a11d\" (UID: \"e845a73a-f851-421e-8708-55e618d3a11d\") " Dec 03 06:30:54 crc kubenswrapper[4818]: I1203 06:30:54.456186 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e845a73a-f851-421e-8708-55e618d3a11d-kube-api-access\") pod \"e845a73a-f851-421e-8708-55e618d3a11d\" (UID: \"e845a73a-f851-421e-8708-55e618d3a11d\") " Dec 03 06:30:54 crc kubenswrapper[4818]: I1203 06:30:54.456894 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e845a73a-f851-421e-8708-55e618d3a11d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e845a73a-f851-421e-8708-55e618d3a11d" (UID: "e845a73a-f851-421e-8708-55e618d3a11d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:30:54 crc kubenswrapper[4818]: I1203 06:30:54.461335 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e845a73a-f851-421e-8708-55e618d3a11d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e845a73a-f851-421e-8708-55e618d3a11d" (UID: "e845a73a-f851-421e-8708-55e618d3a11d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:30:54 crc kubenswrapper[4818]: I1203 06:30:54.558599 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e845a73a-f851-421e-8708-55e618d3a11d-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 06:30:54 crc kubenswrapper[4818]: I1203 06:30:54.558651 4818 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e845a73a-f851-421e-8708-55e618d3a11d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:30:54 crc kubenswrapper[4818]: I1203 06:30:54.975517 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 06:30:54 crc kubenswrapper[4818]: E1203 06:30:54.975715 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e845a73a-f851-421e-8708-55e618d3a11d" containerName="pruner" Dec 03 06:30:54 crc kubenswrapper[4818]: I1203 06:30:54.975727 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e845a73a-f851-421e-8708-55e618d3a11d" containerName="pruner" Dec 03 06:30:54 crc kubenswrapper[4818]: I1203 06:30:54.975852 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="e845a73a-f851-421e-8708-55e618d3a11d" containerName="pruner" Dec 03 06:30:54 crc kubenswrapper[4818]: I1203 06:30:54.976268 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:30:54 crc kubenswrapper[4818]: I1203 06:30:54.983994 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 06:30:55 crc kubenswrapper[4818]: I1203 06:30:55.064617 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-kubelet-dir\") pod \"installer-9-crc\" (UID: \"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:30:55 crc kubenswrapper[4818]: I1203 06:30:55.065273 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-var-lock\") pod \"installer-9-crc\" (UID: \"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:30:55 crc kubenswrapper[4818]: I1203 06:30:55.065317 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-kube-api-access\") pod \"installer-9-crc\" (UID: \"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:30:55 crc kubenswrapper[4818]: I1203 06:30:55.166372 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-kubelet-dir\") pod \"installer-9-crc\" (UID: \"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:30:55 crc kubenswrapper[4818]: I1203 06:30:55.166452 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-var-lock\") pod \"installer-9-crc\" (UID: \"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:30:55 crc kubenswrapper[4818]: I1203 06:30:55.166475 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-kube-api-access\") pod \"installer-9-crc\" (UID: \"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:30:55 crc kubenswrapper[4818]: I1203 06:30:55.166519 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-kubelet-dir\") pod \"installer-9-crc\" (UID: \"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:30:55 crc kubenswrapper[4818]: I1203 06:30:55.166532 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-var-lock\") pod \"installer-9-crc\" (UID: \"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:30:55 crc kubenswrapper[4818]: I1203 06:30:55.199499 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-kube-api-access\") pod \"installer-9-crc\" (UID: \"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:30:55 crc kubenswrapper[4818]: I1203 06:30:55.203491 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e845a73a-f851-421e-8708-55e618d3a11d","Type":"ContainerDied","Data":"c7b11145e3f0872d35322f0ba788189a9e35ab9c55c0727acc2b12f6c6b5c3ba"} Dec 03 06:30:55 crc kubenswrapper[4818]: I1203 06:30:55.203531 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7b11145e3f0872d35322f0ba788189a9e35ab9c55c0727acc2b12f6c6b5c3ba" Dec 03 06:30:55 crc kubenswrapper[4818]: I1203 06:30:55.203594 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 06:30:55 crc kubenswrapper[4818]: I1203 06:30:55.301167 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:30:55 crc kubenswrapper[4818]: I1203 06:30:55.503016 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 06:30:55 crc kubenswrapper[4818]: W1203 06:30:55.518946 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod24a1ab03_bbf6_4d53_9adb_ddc42b90fdde.slice/crio-893d0323fd0ce884b024f493e57892b9ee6c04618dc4356e1491f3f0c691a0e8 WatchSource:0}: Error finding container 893d0323fd0ce884b024f493e57892b9ee6c04618dc4356e1491f3f0c691a0e8: Status 404 returned error can't find the container with id 893d0323fd0ce884b024f493e57892b9ee6c04618dc4356e1491f3f0c691a0e8 Dec 03 06:30:56 crc kubenswrapper[4818]: I1203 06:30:56.208154 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde","Type":"ContainerStarted","Data":"893d0323fd0ce884b024f493e57892b9ee6c04618dc4356e1491f3f0c691a0e8"} Dec 03 06:30:57 crc kubenswrapper[4818]: I1203 06:30:57.219288 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde","Type":"ContainerStarted","Data":"dd6041b5271c0de55f8a9b33927ef692b055606ec14afcf8cb4a5cb94d4b228f"} Dec 03 06:30:57 crc kubenswrapper[4818]: I1203 06:30:57.240754 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.240731777 podStartE2EDuration="3.240731777s" podCreationTimestamp="2025-12-03 06:30:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:30:57.234619965 +0000 UTC m=+214.926228807" watchObservedRunningTime="2025-12-03 06:30:57.240731777 +0000 UTC m=+214.932340559" Dec 03 06:31:05 crc kubenswrapper[4818]: I1203 06:31:05.263734 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7vlw" event={"ID":"75955bfe-ae6e-481d-84b5-3de486a86827","Type":"ContainerStarted","Data":"893aac79870bf08760a3dadb13a02f9f0d3f8d8e4f7bdb9fc0176bd870851cda"} Dec 03 06:31:05 crc kubenswrapper[4818]: I1203 06:31:05.266616 4818 generic.go:334] "Generic (PLEG): container finished" podID="91223086-535a-47c4-a843-3bb100a7da50" containerID="8f0fe5cc139e0b9e63edec1a51c00eac2621dccc305d594aa83e600ea5c4212c" exitCode=0 Dec 03 06:31:05 crc kubenswrapper[4818]: I1203 06:31:05.266673 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-965hr" event={"ID":"91223086-535a-47c4-a843-3bb100a7da50","Type":"ContainerDied","Data":"8f0fe5cc139e0b9e63edec1a51c00eac2621dccc305d594aa83e600ea5c4212c"} Dec 03 06:31:06 crc kubenswrapper[4818]: I1203 06:31:06.275723 4818 generic.go:334] "Generic (PLEG): container finished" podID="75955bfe-ae6e-481d-84b5-3de486a86827" containerID="893aac79870bf08760a3dadb13a02f9f0d3f8d8e4f7bdb9fc0176bd870851cda" exitCode=0 Dec 03 06:31:06 crc kubenswrapper[4818]: I1203 06:31:06.275786 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7vlw" event={"ID":"75955bfe-ae6e-481d-84b5-3de486a86827","Type":"ContainerDied","Data":"893aac79870bf08760a3dadb13a02f9f0d3f8d8e4f7bdb9fc0176bd870851cda"} Dec 03 06:31:07 crc kubenswrapper[4818]: I1203 06:31:07.732256 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fv5lc"] Dec 03 06:31:10 crc kubenswrapper[4818]: I1203 06:31:10.296301 4818 generic.go:334] "Generic (PLEG): container finished" podID="366db0ca-c6f6-49ee-9968-05a60e74dae5" containerID="271af713cad85603742094c16dcf65d4c7889c819f31bbc827e9219152e89e4e" exitCode=0 Dec 03 06:31:10 crc kubenswrapper[4818]: I1203 06:31:10.296377 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrv77" event={"ID":"366db0ca-c6f6-49ee-9968-05a60e74dae5","Type":"ContainerDied","Data":"271af713cad85603742094c16dcf65d4c7889c819f31bbc827e9219152e89e4e"} Dec 03 06:31:10 crc kubenswrapper[4818]: I1203 06:31:10.301782 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fzf6r" event={"ID":"b81e21fd-b361-4853-9a91-e034667fce63","Type":"ContainerStarted","Data":"c3c1e46a5299eaf176609042158815aa85e462015951111370b31be3c5702e2b"} Dec 03 06:31:10 crc kubenswrapper[4818]: I1203 06:31:10.304153 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfzk6" event={"ID":"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d","Type":"ContainerStarted","Data":"53c09a0eaa9fb6d178733be4ec2717ea22fd9939aee991f98e87aefc1e3ab8bc"} Dec 03 06:31:10 crc kubenswrapper[4818]: I1203 06:31:10.307059 4818 generic.go:334] "Generic (PLEG): container finished" podID="46f85cb1-3840-4ff8-a506-d41af5a5f2f2" containerID="488adf5c3b3e30a4239a17553b5741f79d3a9ca3a7c1a53067e507af8282a4aa" exitCode=0 Dec 03 06:31:10 crc kubenswrapper[4818]: I1203 06:31:10.307127 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfvqs" event={"ID":"46f85cb1-3840-4ff8-a506-d41af5a5f2f2","Type":"ContainerDied","Data":"488adf5c3b3e30a4239a17553b5741f79d3a9ca3a7c1a53067e507af8282a4aa"} Dec 03 06:31:10 crc kubenswrapper[4818]: I1203 06:31:10.309859 4818 generic.go:334] "Generic (PLEG): container finished" podID="0c4ffde8-459a-4634-b567-5f900d40ceec" containerID="196fff865cfdab8c5f02bfefaa00f1df7162323673ab198c96326c7a3f8c62db" exitCode=0 Dec 03 06:31:10 crc kubenswrapper[4818]: I1203 06:31:10.309936 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrfsv" event={"ID":"0c4ffde8-459a-4634-b567-5f900d40ceec","Type":"ContainerDied","Data":"196fff865cfdab8c5f02bfefaa00f1df7162323673ab198c96326c7a3f8c62db"} Dec 03 06:31:10 crc kubenswrapper[4818]: I1203 06:31:10.315709 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7vlw" event={"ID":"75955bfe-ae6e-481d-84b5-3de486a86827","Type":"ContainerStarted","Data":"8e606076dd130f5dbe3724a712bad6f18f49e260d7aadf33a5b803edde0431f8"} Dec 03 06:31:10 crc kubenswrapper[4818]: I1203 06:31:10.320299 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-965hr" event={"ID":"91223086-535a-47c4-a843-3bb100a7da50","Type":"ContainerStarted","Data":"3e37f9cc5816052f7c6991762135723f0f86e6f484e32bd8e2870d46593d55bc"} Dec 03 06:31:10 crc kubenswrapper[4818]: I1203 06:31:10.323531 4818 generic.go:334] "Generic (PLEG): container finished" podID="0301f892-ca1d-4f8b-a429-6d0ad96a50dc" containerID="beb806bf0b4a2470eb3fd8b472967f2c9d9a177842e7782b4cbb64047bdf3217" exitCode=0 Dec 03 06:31:10 crc kubenswrapper[4818]: I1203 06:31:10.323575 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7wwb" event={"ID":"0301f892-ca1d-4f8b-a429-6d0ad96a50dc","Type":"ContainerDied","Data":"beb806bf0b4a2470eb3fd8b472967f2c9d9a177842e7782b4cbb64047bdf3217"} Dec 03 06:31:10 crc kubenswrapper[4818]: I1203 06:31:10.422394 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-965hr" podStartSLOduration=2.498382167 podStartE2EDuration="1m2.422375773s" podCreationTimestamp="2025-12-03 06:30:08 +0000 UTC" firstStartedPulling="2025-12-03 06:30:09.658626736 +0000 UTC m=+167.350235488" lastFinishedPulling="2025-12-03 06:31:09.582620332 +0000 UTC m=+227.274229094" observedRunningTime="2025-12-03 06:31:10.420410644 +0000 UTC m=+228.112019396" watchObservedRunningTime="2025-12-03 06:31:10.422375773 +0000 UTC m=+228.113984525" Dec 03 06:31:10 crc kubenswrapper[4818]: I1203 06:31:10.486334 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x7vlw" podStartSLOduration=2.326266393 podStartE2EDuration="1m1.486308801s" podCreationTimestamp="2025-12-03 06:30:09 +0000 UTC" firstStartedPulling="2025-12-03 06:30:10.68772679 +0000 UTC m=+168.379335542" lastFinishedPulling="2025-12-03 06:31:09.847769198 +0000 UTC m=+227.539377950" observedRunningTime="2025-12-03 06:31:10.484156627 +0000 UTC m=+228.175765379" watchObservedRunningTime="2025-12-03 06:31:10.486308801 +0000 UTC m=+228.177917553" Dec 03 06:31:11 crc kubenswrapper[4818]: I1203 06:31:11.330356 4818 generic.go:334] "Generic (PLEG): container finished" podID="ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d" containerID="53c09a0eaa9fb6d178733be4ec2717ea22fd9939aee991f98e87aefc1e3ab8bc" exitCode=0 Dec 03 06:31:11 crc kubenswrapper[4818]: I1203 06:31:11.330399 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfzk6" event={"ID":"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d","Type":"ContainerDied","Data":"53c09a0eaa9fb6d178733be4ec2717ea22fd9939aee991f98e87aefc1e3ab8bc"} Dec 03 06:31:11 crc kubenswrapper[4818]: I1203 06:31:11.336709 4818 generic.go:334] "Generic (PLEG): container finished" podID="b81e21fd-b361-4853-9a91-e034667fce63" containerID="c3c1e46a5299eaf176609042158815aa85e462015951111370b31be3c5702e2b" exitCode=0 Dec 03 06:31:11 crc kubenswrapper[4818]: I1203 06:31:11.337883 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fzf6r" event={"ID":"b81e21fd-b361-4853-9a91-e034667fce63","Type":"ContainerDied","Data":"c3c1e46a5299eaf176609042158815aa85e462015951111370b31be3c5702e2b"} Dec 03 06:31:12 crc kubenswrapper[4818]: I1203 06:31:12.343055 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7wwb" event={"ID":"0301f892-ca1d-4f8b-a429-6d0ad96a50dc","Type":"ContainerStarted","Data":"04caaf0e08ce8e688a22ab83ddfb7237545bbbb93f5e378ad7d0955d1d9446fd"} Dec 03 06:31:12 crc kubenswrapper[4818]: I1203 06:31:12.344856 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrv77" event={"ID":"366db0ca-c6f6-49ee-9968-05a60e74dae5","Type":"ContainerStarted","Data":"7068344c887cd4f87362affeb7c6f0e285196733b2534c03a258017f80678fe9"} Dec 03 06:31:12 crc kubenswrapper[4818]: I1203 06:31:12.346714 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfzk6" event={"ID":"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d","Type":"ContainerStarted","Data":"2e458f4d6f14c6f322ac602b598f597475f94fb237280e5583e4e899f4b0741f"} Dec 03 06:31:12 crc kubenswrapper[4818]: I1203 06:31:12.349262 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fzf6r" event={"ID":"b81e21fd-b361-4853-9a91-e034667fce63","Type":"ContainerStarted","Data":"42c504d96b3da4fc73c2972ff2961633f986514662d15a4ca0b5f8a607e8d338"} Dec 03 06:31:12 crc kubenswrapper[4818]: I1203 06:31:12.351125 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfvqs" event={"ID":"46f85cb1-3840-4ff8-a506-d41af5a5f2f2","Type":"ContainerStarted","Data":"b8f89c6c0885f4d1a74596f40d7c46e4a19f574edcb5ec10b354cc5a9b6b491c"} Dec 03 06:31:12 crc kubenswrapper[4818]: I1203 06:31:12.352894 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrfsv" event={"ID":"0c4ffde8-459a-4634-b567-5f900d40ceec","Type":"ContainerStarted","Data":"1ea4fe4551a71e496e406f333497c69daf15b0f43c41a115de3c6500e2520da6"} Dec 03 06:31:12 crc kubenswrapper[4818]: I1203 06:31:12.362842 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k7wwb" podStartSLOduration=3.609119155 podStartE2EDuration="1m6.362808446s" podCreationTimestamp="2025-12-03 06:30:06 +0000 UTC" firstStartedPulling="2025-12-03 06:30:08.61862598 +0000 UTC m=+166.310234742" lastFinishedPulling="2025-12-03 06:31:11.372315281 +0000 UTC m=+229.063924033" observedRunningTime="2025-12-03 06:31:12.35853252 +0000 UTC m=+230.050141272" watchObservedRunningTime="2025-12-03 06:31:12.362808446 +0000 UTC m=+230.054417198" Dec 03 06:31:12 crc kubenswrapper[4818]: I1203 06:31:12.374932 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fzf6r" podStartSLOduration=2.295102851 podStartE2EDuration="1m2.374911597s" podCreationTimestamp="2025-12-03 06:30:10 +0000 UTC" firstStartedPulling="2025-12-03 06:30:11.771800481 +0000 UTC m=+169.463409233" lastFinishedPulling="2025-12-03 06:31:11.851609237 +0000 UTC m=+229.543217979" observedRunningTime="2025-12-03 06:31:12.373993634 +0000 UTC m=+230.065602396" watchObservedRunningTime="2025-12-03 06:31:12.374911597 +0000 UTC m=+230.066520349" Dec 03 06:31:12 crc kubenswrapper[4818]: I1203 06:31:12.392683 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xfzk6" podStartSLOduration=3.242354428 podStartE2EDuration="1m4.392667388s" podCreationTimestamp="2025-12-03 06:30:08 +0000 UTC" firstStartedPulling="2025-12-03 06:30:10.728179155 +0000 UTC m=+168.419787907" lastFinishedPulling="2025-12-03 06:31:11.878492115 +0000 UTC m=+229.570100867" observedRunningTime="2025-12-03 06:31:12.389917989 +0000 UTC m=+230.081526741" watchObservedRunningTime="2025-12-03 06:31:12.392667388 +0000 UTC m=+230.084276140" Dec 03 06:31:12 crc kubenswrapper[4818]: I1203 06:31:12.407599 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zfvqs" podStartSLOduration=3.524606026 podStartE2EDuration="1m6.407579348s" podCreationTimestamp="2025-12-03 06:30:06 +0000 UTC" firstStartedPulling="2025-12-03 06:30:08.615942183 +0000 UTC m=+166.307550935" lastFinishedPulling="2025-12-03 06:31:11.498915505 +0000 UTC m=+229.190524257" observedRunningTime="2025-12-03 06:31:12.404690756 +0000 UTC m=+230.096299518" watchObservedRunningTime="2025-12-03 06:31:12.407579348 +0000 UTC m=+230.099188100" Dec 03 06:31:12 crc kubenswrapper[4818]: I1203 06:31:12.422635 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jrfsv" podStartSLOduration=3.624891137 podStartE2EDuration="1m6.422618352s" podCreationTimestamp="2025-12-03 06:30:06 +0000 UTC" firstStartedPulling="2025-12-03 06:30:08.604604042 +0000 UTC m=+166.296212794" lastFinishedPulling="2025-12-03 06:31:11.402331257 +0000 UTC m=+229.093940009" observedRunningTime="2025-12-03 06:31:12.420206552 +0000 UTC m=+230.111815314" watchObservedRunningTime="2025-12-03 06:31:12.422618352 +0000 UTC m=+230.114227094" Dec 03 06:31:12 crc kubenswrapper[4818]: I1203 06:31:12.437081 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hrv77" podStartSLOduration=3.513096793 podStartE2EDuration="1m5.437061681s" podCreationTimestamp="2025-12-03 06:30:07 +0000 UTC" firstStartedPulling="2025-12-03 06:30:09.645159631 +0000 UTC m=+167.336768383" lastFinishedPulling="2025-12-03 06:31:11.569124519 +0000 UTC m=+229.260733271" observedRunningTime="2025-12-03 06:31:12.435575804 +0000 UTC m=+230.127184566" watchObservedRunningTime="2025-12-03 06:31:12.437061681 +0000 UTC m=+230.128670433" Dec 03 06:31:13 crc kubenswrapper[4818]: I1203 06:31:13.302017 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:31:13 crc kubenswrapper[4818]: I1203 06:31:13.302085 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:31:13 crc kubenswrapper[4818]: I1203 06:31:13.302131 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:31:13 crc kubenswrapper[4818]: I1203 06:31:13.302760 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f"} pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 06:31:13 crc kubenswrapper[4818]: I1203 06:31:13.302838 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" containerID="cri-o://a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f" gracePeriod=600 Dec 03 06:31:14 crc kubenswrapper[4818]: I1203 06:31:14.365976 4818 generic.go:334] "Generic (PLEG): container finished" podID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerID="a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f" exitCode=0 Dec 03 06:31:14 crc kubenswrapper[4818]: I1203 06:31:14.366074 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerDied","Data":"a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f"} Dec 03 06:31:14 crc kubenswrapper[4818]: I1203 06:31:14.366608 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"1f4c468b483e867be2291edf0efa4884963d305fb4fe62f714e10f36adc23d54"} Dec 03 06:31:16 crc kubenswrapper[4818]: I1203 06:31:16.785894 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k7wwb" Dec 03 06:31:16 crc kubenswrapper[4818]: I1203 06:31:16.786463 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k7wwb" Dec 03 06:31:16 crc kubenswrapper[4818]: I1203 06:31:16.850928 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k7wwb" Dec 03 06:31:17 crc kubenswrapper[4818]: I1203 06:31:17.280495 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zfvqs" Dec 03 06:31:17 crc kubenswrapper[4818]: I1203 06:31:17.281145 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jrfsv" Dec 03 06:31:17 crc kubenswrapper[4818]: I1203 06:31:17.281220 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zfvqs" Dec 03 06:31:17 crc kubenswrapper[4818]: I1203 06:31:17.281242 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jrfsv" Dec 03 06:31:17 crc kubenswrapper[4818]: I1203 06:31:17.333738 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jrfsv" Dec 03 06:31:17 crc kubenswrapper[4818]: I1203 06:31:17.349636 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zfvqs" Dec 03 06:31:17 crc kubenswrapper[4818]: I1203 06:31:17.396933 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hrv77" Dec 03 06:31:17 crc kubenswrapper[4818]: I1203 06:31:17.396985 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hrv77" Dec 03 06:31:17 crc kubenswrapper[4818]: I1203 06:31:17.422376 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k7wwb" Dec 03 06:31:17 crc kubenswrapper[4818]: I1203 06:31:17.422660 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jrfsv" Dec 03 06:31:17 crc kubenswrapper[4818]: I1203 06:31:17.427396 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zfvqs" Dec 03 06:31:17 crc kubenswrapper[4818]: I1203 06:31:17.438401 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hrv77" Dec 03 06:31:18 crc kubenswrapper[4818]: I1203 06:31:18.435270 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hrv77" Dec 03 06:31:18 crc kubenswrapper[4818]: I1203 06:31:18.830421 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-965hr" Dec 03 06:31:18 crc kubenswrapper[4818]: I1203 06:31:18.830487 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-965hr" Dec 03 06:31:18 crc kubenswrapper[4818]: I1203 06:31:18.867577 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-965hr" Dec 03 06:31:19 crc kubenswrapper[4818]: I1203 06:31:19.275711 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xfzk6" Dec 03 06:31:19 crc kubenswrapper[4818]: I1203 06:31:19.275966 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xfzk6" Dec 03 06:31:19 crc kubenswrapper[4818]: I1203 06:31:19.320436 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xfzk6" Dec 03 06:31:19 crc kubenswrapper[4818]: I1203 06:31:19.443961 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xfzk6" Dec 03 06:31:19 crc kubenswrapper[4818]: I1203 06:31:19.447526 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-965hr" Dec 03 06:31:19 crc kubenswrapper[4818]: I1203 06:31:19.572767 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zfvqs"] Dec 03 06:31:20 crc kubenswrapper[4818]: I1203 06:31:20.028793 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x7vlw" Dec 03 06:31:20 crc kubenswrapper[4818]: I1203 06:31:20.028942 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x7vlw" Dec 03 06:31:20 crc kubenswrapper[4818]: I1203 06:31:20.087021 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x7vlw" Dec 03 06:31:20 crc kubenswrapper[4818]: I1203 06:31:20.404127 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zfvqs" podUID="46f85cb1-3840-4ff8-a506-d41af5a5f2f2" containerName="registry-server" containerID="cri-o://b8f89c6c0885f4d1a74596f40d7c46e4a19f574edcb5ec10b354cc5a9b6b491c" gracePeriod=2 Dec 03 06:31:20 crc kubenswrapper[4818]: I1203 06:31:20.431430 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fzf6r" Dec 03 06:31:20 crc kubenswrapper[4818]: I1203 06:31:20.432129 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fzf6r" Dec 03 06:31:20 crc kubenswrapper[4818]: I1203 06:31:20.440959 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x7vlw" Dec 03 06:31:20 crc kubenswrapper[4818]: I1203 06:31:20.494536 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fzf6r" Dec 03 06:31:20 crc kubenswrapper[4818]: I1203 06:31:20.566143 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hrv77"] Dec 03 06:31:20 crc kubenswrapper[4818]: I1203 06:31:20.566389 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hrv77" podUID="366db0ca-c6f6-49ee-9968-05a60e74dae5" containerName="registry-server" containerID="cri-o://7068344c887cd4f87362affeb7c6f0e285196733b2534c03a258017f80678fe9" gracePeriod=2 Dec 03 06:31:21 crc kubenswrapper[4818]: I1203 06:31:21.409523 4818 generic.go:334] "Generic (PLEG): container finished" podID="46f85cb1-3840-4ff8-a506-d41af5a5f2f2" containerID="b8f89c6c0885f4d1a74596f40d7c46e4a19f574edcb5ec10b354cc5a9b6b491c" exitCode=0 Dec 03 06:31:21 crc kubenswrapper[4818]: I1203 06:31:21.409850 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfvqs" event={"ID":"46f85cb1-3840-4ff8-a506-d41af5a5f2f2","Type":"ContainerDied","Data":"b8f89c6c0885f4d1a74596f40d7c46e4a19f574edcb5ec10b354cc5a9b6b491c"} Dec 03 06:31:21 crc kubenswrapper[4818]: I1203 06:31:21.447571 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fzf6r" Dec 03 06:31:21 crc kubenswrapper[4818]: I1203 06:31:21.969061 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfzk6"] Dec 03 06:31:22 crc kubenswrapper[4818]: I1203 06:31:22.420669 4818 generic.go:334] "Generic (PLEG): container finished" podID="366db0ca-c6f6-49ee-9968-05a60e74dae5" containerID="7068344c887cd4f87362affeb7c6f0e285196733b2534c03a258017f80678fe9" exitCode=0 Dec 03 06:31:22 crc kubenswrapper[4818]: I1203 06:31:22.420932 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrv77" event={"ID":"366db0ca-c6f6-49ee-9968-05a60e74dae5","Type":"ContainerDied","Data":"7068344c887cd4f87362affeb7c6f0e285196733b2534c03a258017f80678fe9"} Dec 03 06:31:22 crc kubenswrapper[4818]: I1203 06:31:22.421056 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xfzk6" podUID="ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d" containerName="registry-server" containerID="cri-o://2e458f4d6f14c6f322ac602b598f597475f94fb237280e5583e4e899f4b0741f" gracePeriod=2 Dec 03 06:31:22 crc kubenswrapper[4818]: I1203 06:31:22.801911 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfvqs" Dec 03 06:31:22 crc kubenswrapper[4818]: I1203 06:31:22.955519 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2jx7\" (UniqueName: \"kubernetes.io/projected/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-kube-api-access-s2jx7\") pod \"46f85cb1-3840-4ff8-a506-d41af5a5f2f2\" (UID: \"46f85cb1-3840-4ff8-a506-d41af5a5f2f2\") " Dec 03 06:31:22 crc kubenswrapper[4818]: I1203 06:31:22.955650 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-catalog-content\") pod \"46f85cb1-3840-4ff8-a506-d41af5a5f2f2\" (UID: \"46f85cb1-3840-4ff8-a506-d41af5a5f2f2\") " Dec 03 06:31:22 crc kubenswrapper[4818]: I1203 06:31:22.955678 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-utilities\") pod \"46f85cb1-3840-4ff8-a506-d41af5a5f2f2\" (UID: \"46f85cb1-3840-4ff8-a506-d41af5a5f2f2\") " Dec 03 06:31:22 crc kubenswrapper[4818]: I1203 06:31:22.956911 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-utilities" (OuterVolumeSpecName: "utilities") pod "46f85cb1-3840-4ff8-a506-d41af5a5f2f2" (UID: "46f85cb1-3840-4ff8-a506-d41af5a5f2f2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:31:22 crc kubenswrapper[4818]: I1203 06:31:22.962612 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-kube-api-access-s2jx7" (OuterVolumeSpecName: "kube-api-access-s2jx7") pod "46f85cb1-3840-4ff8-a506-d41af5a5f2f2" (UID: "46f85cb1-3840-4ff8-a506-d41af5a5f2f2"). InnerVolumeSpecName "kube-api-access-s2jx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.024769 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46f85cb1-3840-4ff8-a506-d41af5a5f2f2" (UID: "46f85cb1-3840-4ff8-a506-d41af5a5f2f2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.057874 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.057909 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.057923 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2jx7\" (UniqueName: \"kubernetes.io/projected/46f85cb1-3840-4ff8-a506-d41af5a5f2f2-kube-api-access-s2jx7\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.165154 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hrv77" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.259798 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/366db0ca-c6f6-49ee-9968-05a60e74dae5-utilities\") pod \"366db0ca-c6f6-49ee-9968-05a60e74dae5\" (UID: \"366db0ca-c6f6-49ee-9968-05a60e74dae5\") " Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.259892 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/366db0ca-c6f6-49ee-9968-05a60e74dae5-catalog-content\") pod \"366db0ca-c6f6-49ee-9968-05a60e74dae5\" (UID: \"366db0ca-c6f6-49ee-9968-05a60e74dae5\") " Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.260001 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc7nz\" (UniqueName: \"kubernetes.io/projected/366db0ca-c6f6-49ee-9968-05a60e74dae5-kube-api-access-qc7nz\") pod \"366db0ca-c6f6-49ee-9968-05a60e74dae5\" (UID: \"366db0ca-c6f6-49ee-9968-05a60e74dae5\") " Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.261232 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/366db0ca-c6f6-49ee-9968-05a60e74dae5-utilities" (OuterVolumeSpecName: "utilities") pod "366db0ca-c6f6-49ee-9968-05a60e74dae5" (UID: "366db0ca-c6f6-49ee-9968-05a60e74dae5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.267869 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/366db0ca-c6f6-49ee-9968-05a60e74dae5-kube-api-access-qc7nz" (OuterVolumeSpecName: "kube-api-access-qc7nz") pod "366db0ca-c6f6-49ee-9968-05a60e74dae5" (UID: "366db0ca-c6f6-49ee-9968-05a60e74dae5"). InnerVolumeSpecName "kube-api-access-qc7nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.299858 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfzk6" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.351338 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/366db0ca-c6f6-49ee-9968-05a60e74dae5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "366db0ca-c6f6-49ee-9968-05a60e74dae5" (UID: "366db0ca-c6f6-49ee-9968-05a60e74dae5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.361546 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-catalog-content\") pod \"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d\" (UID: \"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d\") " Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.361595 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdt47\" (UniqueName: \"kubernetes.io/projected/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-kube-api-access-bdt47\") pod \"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d\" (UID: \"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d\") " Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.361672 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-utilities\") pod \"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d\" (UID: \"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d\") " Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.361906 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc7nz\" (UniqueName: \"kubernetes.io/projected/366db0ca-c6f6-49ee-9968-05a60e74dae5-kube-api-access-qc7nz\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.361923 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/366db0ca-c6f6-49ee-9968-05a60e74dae5-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.361932 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/366db0ca-c6f6-49ee-9968-05a60e74dae5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.362850 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-utilities" (OuterVolumeSpecName: "utilities") pod "ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d" (UID: "ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.366581 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-kube-api-access-bdt47" (OuterVolumeSpecName: "kube-api-access-bdt47") pod "ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d" (UID: "ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d"). InnerVolumeSpecName "kube-api-access-bdt47". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.379385 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d" (UID: "ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.427917 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hrv77" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.427922 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hrv77" event={"ID":"366db0ca-c6f6-49ee-9968-05a60e74dae5","Type":"ContainerDied","Data":"aaed7d315eefa0f5c38cf0687aee9fa92846f337bf0557ddc7adfb90604dcdb0"} Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.428065 4818 scope.go:117] "RemoveContainer" containerID="7068344c887cd4f87362affeb7c6f0e285196733b2534c03a258017f80678fe9" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.432931 4818 generic.go:334] "Generic (PLEG): container finished" podID="ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d" containerID="2e458f4d6f14c6f322ac602b598f597475f94fb237280e5583e4e899f4b0741f" exitCode=0 Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.432999 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfzk6" event={"ID":"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d","Type":"ContainerDied","Data":"2e458f4d6f14c6f322ac602b598f597475f94fb237280e5583e4e899f4b0741f"} Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.433027 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xfzk6" event={"ID":"ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d","Type":"ContainerDied","Data":"ba57498bff52ce47e338363a843becdd6295fcf2ec3ca636dbb8fe453930e5d2"} Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.433094 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xfzk6" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.442698 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zfvqs" event={"ID":"46f85cb1-3840-4ff8-a506-d41af5a5f2f2","Type":"ContainerDied","Data":"ae072f843b0ba8de6c4af5338aa15e157a03f03e1abe0155846faf6051e3c6f4"} Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.442743 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zfvqs" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.450690 4818 scope.go:117] "RemoveContainer" containerID="271af713cad85603742094c16dcf65d4c7889c819f31bbc827e9219152e89e4e" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.463677 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.463712 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.463725 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdt47\" (UniqueName: \"kubernetes.io/projected/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d-kube-api-access-bdt47\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.466446 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hrv77"] Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.469188 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hrv77"] Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.483887 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfzk6"] Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.488169 4818 scope.go:117] "RemoveContainer" containerID="7caa155be3fa4a7bc0fa4329fcb2e2f4b74412181d26ee1ba285e65a7d136de5" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.489005 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xfzk6"] Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.497734 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zfvqs"] Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.501972 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zfvqs"] Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.510740 4818 scope.go:117] "RemoveContainer" containerID="2e458f4d6f14c6f322ac602b598f597475f94fb237280e5583e4e899f4b0741f" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.527237 4818 scope.go:117] "RemoveContainer" containerID="53c09a0eaa9fb6d178733be4ec2717ea22fd9939aee991f98e87aefc1e3ab8bc" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.576210 4818 scope.go:117] "RemoveContainer" containerID="8345a9f57c7efc24bd7ff9af58ee1afb3e2fc140acab644908909991f7f9fc8a" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.598328 4818 scope.go:117] "RemoveContainer" containerID="2e458f4d6f14c6f322ac602b598f597475f94fb237280e5583e4e899f4b0741f" Dec 03 06:31:23 crc kubenswrapper[4818]: E1203 06:31:23.598734 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e458f4d6f14c6f322ac602b598f597475f94fb237280e5583e4e899f4b0741f\": container with ID starting with 2e458f4d6f14c6f322ac602b598f597475f94fb237280e5583e4e899f4b0741f not found: ID does not exist" containerID="2e458f4d6f14c6f322ac602b598f597475f94fb237280e5583e4e899f4b0741f" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.598766 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e458f4d6f14c6f322ac602b598f597475f94fb237280e5583e4e899f4b0741f"} err="failed to get container status \"2e458f4d6f14c6f322ac602b598f597475f94fb237280e5583e4e899f4b0741f\": rpc error: code = NotFound desc = could not find container \"2e458f4d6f14c6f322ac602b598f597475f94fb237280e5583e4e899f4b0741f\": container with ID starting with 2e458f4d6f14c6f322ac602b598f597475f94fb237280e5583e4e899f4b0741f not found: ID does not exist" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.598791 4818 scope.go:117] "RemoveContainer" containerID="53c09a0eaa9fb6d178733be4ec2717ea22fd9939aee991f98e87aefc1e3ab8bc" Dec 03 06:31:23 crc kubenswrapper[4818]: E1203 06:31:23.599111 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53c09a0eaa9fb6d178733be4ec2717ea22fd9939aee991f98e87aefc1e3ab8bc\": container with ID starting with 53c09a0eaa9fb6d178733be4ec2717ea22fd9939aee991f98e87aefc1e3ab8bc not found: ID does not exist" containerID="53c09a0eaa9fb6d178733be4ec2717ea22fd9939aee991f98e87aefc1e3ab8bc" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.599242 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53c09a0eaa9fb6d178733be4ec2717ea22fd9939aee991f98e87aefc1e3ab8bc"} err="failed to get container status \"53c09a0eaa9fb6d178733be4ec2717ea22fd9939aee991f98e87aefc1e3ab8bc\": rpc error: code = NotFound desc = could not find container \"53c09a0eaa9fb6d178733be4ec2717ea22fd9939aee991f98e87aefc1e3ab8bc\": container with ID starting with 53c09a0eaa9fb6d178733be4ec2717ea22fd9939aee991f98e87aefc1e3ab8bc not found: ID does not exist" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.599365 4818 scope.go:117] "RemoveContainer" containerID="8345a9f57c7efc24bd7ff9af58ee1afb3e2fc140acab644908909991f7f9fc8a" Dec 03 06:31:23 crc kubenswrapper[4818]: E1203 06:31:23.599955 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8345a9f57c7efc24bd7ff9af58ee1afb3e2fc140acab644908909991f7f9fc8a\": container with ID starting with 8345a9f57c7efc24bd7ff9af58ee1afb3e2fc140acab644908909991f7f9fc8a not found: ID does not exist" containerID="8345a9f57c7efc24bd7ff9af58ee1afb3e2fc140acab644908909991f7f9fc8a" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.600003 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8345a9f57c7efc24bd7ff9af58ee1afb3e2fc140acab644908909991f7f9fc8a"} err="failed to get container status \"8345a9f57c7efc24bd7ff9af58ee1afb3e2fc140acab644908909991f7f9fc8a\": rpc error: code = NotFound desc = could not find container \"8345a9f57c7efc24bd7ff9af58ee1afb3e2fc140acab644908909991f7f9fc8a\": container with ID starting with 8345a9f57c7efc24bd7ff9af58ee1afb3e2fc140acab644908909991f7f9fc8a not found: ID does not exist" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.600040 4818 scope.go:117] "RemoveContainer" containerID="b8f89c6c0885f4d1a74596f40d7c46e4a19f574edcb5ec10b354cc5a9b6b491c" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.614030 4818 scope.go:117] "RemoveContainer" containerID="488adf5c3b3e30a4239a17553b5741f79d3a9ca3a7c1a53067e507af8282a4aa" Dec 03 06:31:23 crc kubenswrapper[4818]: I1203 06:31:23.633706 4818 scope.go:117] "RemoveContainer" containerID="4b91fc3922ad91740eaaaed9e1c9bdb4bee8d9ae7028f086be79ee1f75691376" Dec 03 06:31:24 crc kubenswrapper[4818]: I1203 06:31:24.373861 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fzf6r"] Dec 03 06:31:24 crc kubenswrapper[4818]: I1203 06:31:24.450566 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fzf6r" podUID="b81e21fd-b361-4853-9a91-e034667fce63" containerName="registry-server" containerID="cri-o://42c504d96b3da4fc73c2972ff2961633f986514662d15a4ca0b5f8a607e8d338" gracePeriod=2 Dec 03 06:31:24 crc kubenswrapper[4818]: I1203 06:31:24.752506 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="366db0ca-c6f6-49ee-9968-05a60e74dae5" path="/var/lib/kubelet/pods/366db0ca-c6f6-49ee-9968-05a60e74dae5/volumes" Dec 03 06:31:24 crc kubenswrapper[4818]: I1203 06:31:24.754035 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46f85cb1-3840-4ff8-a506-d41af5a5f2f2" path="/var/lib/kubelet/pods/46f85cb1-3840-4ff8-a506-d41af5a5f2f2/volumes" Dec 03 06:31:24 crc kubenswrapper[4818]: I1203 06:31:24.755254 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d" path="/var/lib/kubelet/pods/ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d/volumes" Dec 03 06:31:24 crc kubenswrapper[4818]: I1203 06:31:24.833948 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fzf6r" Dec 03 06:31:24 crc kubenswrapper[4818]: I1203 06:31:24.990236 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpgzf\" (UniqueName: \"kubernetes.io/projected/b81e21fd-b361-4853-9a91-e034667fce63-kube-api-access-vpgzf\") pod \"b81e21fd-b361-4853-9a91-e034667fce63\" (UID: \"b81e21fd-b361-4853-9a91-e034667fce63\") " Dec 03 06:31:24 crc kubenswrapper[4818]: I1203 06:31:24.990857 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b81e21fd-b361-4853-9a91-e034667fce63-utilities\") pod \"b81e21fd-b361-4853-9a91-e034667fce63\" (UID: \"b81e21fd-b361-4853-9a91-e034667fce63\") " Dec 03 06:31:24 crc kubenswrapper[4818]: I1203 06:31:24.991753 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b81e21fd-b361-4853-9a91-e034667fce63-utilities" (OuterVolumeSpecName: "utilities") pod "b81e21fd-b361-4853-9a91-e034667fce63" (UID: "b81e21fd-b361-4853-9a91-e034667fce63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:31:24 crc kubenswrapper[4818]: I1203 06:31:24.991754 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b81e21fd-b361-4853-9a91-e034667fce63-catalog-content\") pod \"b81e21fd-b361-4853-9a91-e034667fce63\" (UID: \"b81e21fd-b361-4853-9a91-e034667fce63\") " Dec 03 06:31:24 crc kubenswrapper[4818]: I1203 06:31:24.992292 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b81e21fd-b361-4853-9a91-e034667fce63-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:24 crc kubenswrapper[4818]: I1203 06:31:24.996610 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b81e21fd-b361-4853-9a91-e034667fce63-kube-api-access-vpgzf" (OuterVolumeSpecName: "kube-api-access-vpgzf") pod "b81e21fd-b361-4853-9a91-e034667fce63" (UID: "b81e21fd-b361-4853-9a91-e034667fce63"). InnerVolumeSpecName "kube-api-access-vpgzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.093744 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpgzf\" (UniqueName: \"kubernetes.io/projected/b81e21fd-b361-4853-9a91-e034667fce63-kube-api-access-vpgzf\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.118375 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b81e21fd-b361-4853-9a91-e034667fce63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b81e21fd-b361-4853-9a91-e034667fce63" (UID: "b81e21fd-b361-4853-9a91-e034667fce63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.194358 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b81e21fd-b361-4853-9a91-e034667fce63-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.460034 4818 generic.go:334] "Generic (PLEG): container finished" podID="b81e21fd-b361-4853-9a91-e034667fce63" containerID="42c504d96b3da4fc73c2972ff2961633f986514662d15a4ca0b5f8a607e8d338" exitCode=0 Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.460113 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fzf6r" event={"ID":"b81e21fd-b361-4853-9a91-e034667fce63","Type":"ContainerDied","Data":"42c504d96b3da4fc73c2972ff2961633f986514662d15a4ca0b5f8a607e8d338"} Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.460197 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fzf6r" Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.460225 4818 scope.go:117] "RemoveContainer" containerID="42c504d96b3da4fc73c2972ff2961633f986514662d15a4ca0b5f8a607e8d338" Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.460206 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fzf6r" event={"ID":"b81e21fd-b361-4853-9a91-e034667fce63","Type":"ContainerDied","Data":"217dfb9b5dc13e3b154236214528fe2fb8085b020b3110f7a2e5380419bd1e9d"} Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.484492 4818 scope.go:117] "RemoveContainer" containerID="c3c1e46a5299eaf176609042158815aa85e462015951111370b31be3c5702e2b" Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.502110 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fzf6r"] Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.505667 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fzf6r"] Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.514726 4818 scope.go:117] "RemoveContainer" containerID="879661c34ad5ece5a71069376c8adc67fea3f6dbc8d6e5dc4559259ee15a180d" Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.546347 4818 scope.go:117] "RemoveContainer" containerID="42c504d96b3da4fc73c2972ff2961633f986514662d15a4ca0b5f8a607e8d338" Dec 03 06:31:25 crc kubenswrapper[4818]: E1203 06:31:25.547013 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42c504d96b3da4fc73c2972ff2961633f986514662d15a4ca0b5f8a607e8d338\": container with ID starting with 42c504d96b3da4fc73c2972ff2961633f986514662d15a4ca0b5f8a607e8d338 not found: ID does not exist" containerID="42c504d96b3da4fc73c2972ff2961633f986514662d15a4ca0b5f8a607e8d338" Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.547075 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42c504d96b3da4fc73c2972ff2961633f986514662d15a4ca0b5f8a607e8d338"} err="failed to get container status \"42c504d96b3da4fc73c2972ff2961633f986514662d15a4ca0b5f8a607e8d338\": rpc error: code = NotFound desc = could not find container \"42c504d96b3da4fc73c2972ff2961633f986514662d15a4ca0b5f8a607e8d338\": container with ID starting with 42c504d96b3da4fc73c2972ff2961633f986514662d15a4ca0b5f8a607e8d338 not found: ID does not exist" Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.547099 4818 scope.go:117] "RemoveContainer" containerID="c3c1e46a5299eaf176609042158815aa85e462015951111370b31be3c5702e2b" Dec 03 06:31:25 crc kubenswrapper[4818]: E1203 06:31:25.547597 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3c1e46a5299eaf176609042158815aa85e462015951111370b31be3c5702e2b\": container with ID starting with c3c1e46a5299eaf176609042158815aa85e462015951111370b31be3c5702e2b not found: ID does not exist" containerID="c3c1e46a5299eaf176609042158815aa85e462015951111370b31be3c5702e2b" Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.547620 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3c1e46a5299eaf176609042158815aa85e462015951111370b31be3c5702e2b"} err="failed to get container status \"c3c1e46a5299eaf176609042158815aa85e462015951111370b31be3c5702e2b\": rpc error: code = NotFound desc = could not find container \"c3c1e46a5299eaf176609042158815aa85e462015951111370b31be3c5702e2b\": container with ID starting with c3c1e46a5299eaf176609042158815aa85e462015951111370b31be3c5702e2b not found: ID does not exist" Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.547649 4818 scope.go:117] "RemoveContainer" containerID="879661c34ad5ece5a71069376c8adc67fea3f6dbc8d6e5dc4559259ee15a180d" Dec 03 06:31:25 crc kubenswrapper[4818]: E1203 06:31:25.548072 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"879661c34ad5ece5a71069376c8adc67fea3f6dbc8d6e5dc4559259ee15a180d\": container with ID starting with 879661c34ad5ece5a71069376c8adc67fea3f6dbc8d6e5dc4559259ee15a180d not found: ID does not exist" containerID="879661c34ad5ece5a71069376c8adc67fea3f6dbc8d6e5dc4559259ee15a180d" Dec 03 06:31:25 crc kubenswrapper[4818]: I1203 06:31:25.548095 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"879661c34ad5ece5a71069376c8adc67fea3f6dbc8d6e5dc4559259ee15a180d"} err="failed to get container status \"879661c34ad5ece5a71069376c8adc67fea3f6dbc8d6e5dc4559259ee15a180d\": rpc error: code = NotFound desc = could not find container \"879661c34ad5ece5a71069376c8adc67fea3f6dbc8d6e5dc4559259ee15a180d\": container with ID starting with 879661c34ad5ece5a71069376c8adc67fea3f6dbc8d6e5dc4559259ee15a180d not found: ID does not exist" Dec 03 06:31:26 crc kubenswrapper[4818]: I1203 06:31:26.746048 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b81e21fd-b361-4853-9a91-e034667fce63" path="/var/lib/kubelet/pods/b81e21fd-b361-4853-9a91-e034667fce63/volumes" Dec 03 06:31:32 crc kubenswrapper[4818]: I1203 06:31:32.763879 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" podUID="d7f12c1b-f504-4155-bd29-bcb6ab57b646" containerName="oauth-openshift" containerID="cri-o://8cca0fe7a690cbf136550efe0435f181a102b59974fe93863f2476a130111250" gracePeriod=15 Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.204744 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.237252 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-57866998d-v97sw"] Dec 03 06:31:33 crc kubenswrapper[4818]: E1203 06:31:33.237502 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="366db0ca-c6f6-49ee-9968-05a60e74dae5" containerName="extract-utilities" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.237515 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="366db0ca-c6f6-49ee-9968-05a60e74dae5" containerName="extract-utilities" Dec 03 06:31:33 crc kubenswrapper[4818]: E1203 06:31:33.237529 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="366db0ca-c6f6-49ee-9968-05a60e74dae5" containerName="registry-server" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.237537 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="366db0ca-c6f6-49ee-9968-05a60e74dae5" containerName="registry-server" Dec 03 06:31:33 crc kubenswrapper[4818]: E1203 06:31:33.237551 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b81e21fd-b361-4853-9a91-e034667fce63" containerName="extract-utilities" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.237560 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b81e21fd-b361-4853-9a91-e034667fce63" containerName="extract-utilities" Dec 03 06:31:33 crc kubenswrapper[4818]: E1203 06:31:33.237573 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b81e21fd-b361-4853-9a91-e034667fce63" containerName="registry-server" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.237580 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b81e21fd-b361-4853-9a91-e034667fce63" containerName="registry-server" Dec 03 06:31:33 crc kubenswrapper[4818]: E1203 06:31:33.237591 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d" containerName="registry-server" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.237599 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d" containerName="registry-server" Dec 03 06:31:33 crc kubenswrapper[4818]: E1203 06:31:33.237615 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f85cb1-3840-4ff8-a506-d41af5a5f2f2" containerName="registry-server" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.237624 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f85cb1-3840-4ff8-a506-d41af5a5f2f2" containerName="registry-server" Dec 03 06:31:33 crc kubenswrapper[4818]: E1203 06:31:33.237640 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d" containerName="extract-content" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.237651 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d" containerName="extract-content" Dec 03 06:31:33 crc kubenswrapper[4818]: E1203 06:31:33.237666 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="366db0ca-c6f6-49ee-9968-05a60e74dae5" containerName="extract-content" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.237679 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="366db0ca-c6f6-49ee-9968-05a60e74dae5" containerName="extract-content" Dec 03 06:31:33 crc kubenswrapper[4818]: E1203 06:31:33.237689 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f85cb1-3840-4ff8-a506-d41af5a5f2f2" containerName="extract-utilities" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.237696 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f85cb1-3840-4ff8-a506-d41af5a5f2f2" containerName="extract-utilities" Dec 03 06:31:33 crc kubenswrapper[4818]: E1203 06:31:33.237708 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7f12c1b-f504-4155-bd29-bcb6ab57b646" containerName="oauth-openshift" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.237716 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7f12c1b-f504-4155-bd29-bcb6ab57b646" containerName="oauth-openshift" Dec 03 06:31:33 crc kubenswrapper[4818]: E1203 06:31:33.237728 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d" containerName="extract-utilities" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.237740 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d" containerName="extract-utilities" Dec 03 06:31:33 crc kubenswrapper[4818]: E1203 06:31:33.237757 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b81e21fd-b361-4853-9a91-e034667fce63" containerName="extract-content" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.237768 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b81e21fd-b361-4853-9a91-e034667fce63" containerName="extract-content" Dec 03 06:31:33 crc kubenswrapper[4818]: E1203 06:31:33.237781 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f85cb1-3840-4ff8-a506-d41af5a5f2f2" containerName="extract-content" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.237788 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f85cb1-3840-4ff8-a506-d41af5a5f2f2" containerName="extract-content" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.237982 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f85cb1-3840-4ff8-a506-d41af5a5f2f2" containerName="registry-server" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.238002 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7f12c1b-f504-4155-bd29-bcb6ab57b646" containerName="oauth-openshift" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.238019 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="b81e21fd-b361-4853-9a91-e034667fce63" containerName="registry-server" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.238032 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae7bd5bd-4cac-44e9-a4c6-da6de50a1c8d" containerName="registry-server" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.238048 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="366db0ca-c6f6-49ee-9968-05a60e74dae5" containerName="registry-server" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.238553 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.251169 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57866998d-v97sw"] Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.331233 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-trusted-ca-bundle\") pod \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.331315 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-cliconfig\") pod \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.331378 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-audit-policies\") pod \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.331442 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-session\") pod \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.331508 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-router-certs\") pod \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.331550 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-service-ca\") pod \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.331571 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d7f12c1b-f504-4155-bd29-bcb6ab57b646-audit-dir\") pod \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.331597 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-login\") pod \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.331622 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvlzn\" (UniqueName: \"kubernetes.io/projected/d7f12c1b-f504-4155-bd29-bcb6ab57b646-kube-api-access-pvlzn\") pod \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.331654 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-error\") pod \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.331709 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-ocp-branding-template\") pod \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.331742 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-provider-selection\") pod \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.331766 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-idp-0-file-data\") pod \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.331870 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-serving-cert\") pod \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\" (UID: \"d7f12c1b-f504-4155-bd29-bcb6ab57b646\") " Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.331993 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-session\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.332016 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d7f12c1b-f504-4155-bd29-bcb6ab57b646-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "d7f12c1b-f504-4155-bd29-bcb6ab57b646" (UID: "d7f12c1b-f504-4155-bd29-bcb6ab57b646"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.332050 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.332289 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.332326 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-user-template-error\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.332358 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6d73fd02-ff75-4664-a80e-a7413dda55d8-audit-policies\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.332408 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.332445 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.332491 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-user-template-login\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.332515 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6d73fd02-ff75-4664-a80e-a7413dda55d8-audit-dir\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.332539 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-service-ca\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.332678 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-router-certs\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.332751 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq64n\" (UniqueName: \"kubernetes.io/projected/6d73fd02-ff75-4664-a80e-a7413dda55d8-kube-api-access-sq64n\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.332790 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.332847 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.332863 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "d7f12c1b-f504-4155-bd29-bcb6ab57b646" (UID: "d7f12c1b-f504-4155-bd29-bcb6ab57b646"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.332898 4818 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d7f12c1b-f504-4155-bd29-bcb6ab57b646-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.332961 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "d7f12c1b-f504-4155-bd29-bcb6ab57b646" (UID: "d7f12c1b-f504-4155-bd29-bcb6ab57b646"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.333421 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "d7f12c1b-f504-4155-bd29-bcb6ab57b646" (UID: "d7f12c1b-f504-4155-bd29-bcb6ab57b646"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.333450 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "d7f12c1b-f504-4155-bd29-bcb6ab57b646" (UID: "d7f12c1b-f504-4155-bd29-bcb6ab57b646"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.338149 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "d7f12c1b-f504-4155-bd29-bcb6ab57b646" (UID: "d7f12c1b-f504-4155-bd29-bcb6ab57b646"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.338223 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7f12c1b-f504-4155-bd29-bcb6ab57b646-kube-api-access-pvlzn" (OuterVolumeSpecName: "kube-api-access-pvlzn") pod "d7f12c1b-f504-4155-bd29-bcb6ab57b646" (UID: "d7f12c1b-f504-4155-bd29-bcb6ab57b646"). InnerVolumeSpecName "kube-api-access-pvlzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.338749 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "d7f12c1b-f504-4155-bd29-bcb6ab57b646" (UID: "d7f12c1b-f504-4155-bd29-bcb6ab57b646"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.339013 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "d7f12c1b-f504-4155-bd29-bcb6ab57b646" (UID: "d7f12c1b-f504-4155-bd29-bcb6ab57b646"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.339332 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "d7f12c1b-f504-4155-bd29-bcb6ab57b646" (UID: "d7f12c1b-f504-4155-bd29-bcb6ab57b646"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.339745 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "d7f12c1b-f504-4155-bd29-bcb6ab57b646" (UID: "d7f12c1b-f504-4155-bd29-bcb6ab57b646"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.344199 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "d7f12c1b-f504-4155-bd29-bcb6ab57b646" (UID: "d7f12c1b-f504-4155-bd29-bcb6ab57b646"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.344221 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "d7f12c1b-f504-4155-bd29-bcb6ab57b646" (UID: "d7f12c1b-f504-4155-bd29-bcb6ab57b646"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.344549 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "d7f12c1b-f504-4155-bd29-bcb6ab57b646" (UID: "d7f12c1b-f504-4155-bd29-bcb6ab57b646"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.434672 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq64n\" (UniqueName: \"kubernetes.io/projected/6d73fd02-ff75-4664-a80e-a7413dda55d8-kube-api-access-sq64n\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.434802 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.434915 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.434983 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-session\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.435028 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.435084 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.435138 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-user-template-error\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.435222 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6d73fd02-ff75-4664-a80e-a7413dda55d8-audit-policies\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.435354 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.435425 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.435501 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-user-template-login\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.435545 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6d73fd02-ff75-4664-a80e-a7413dda55d8-audit-dir\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.435594 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-service-ca\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.435641 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-router-certs\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.435790 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvlzn\" (UniqueName: \"kubernetes.io/projected/d7f12c1b-f504-4155-bd29-bcb6ab57b646-kube-api-access-pvlzn\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.435860 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.435892 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.435924 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.435952 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.435981 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.436007 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.436038 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.436064 4818 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.436090 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.436120 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.436146 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.436172 4818 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d7f12c1b-f504-4155-bd29-bcb6ab57b646-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.436523 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6d73fd02-ff75-4664-a80e-a7413dda55d8-audit-dir\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.436698 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.437483 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.437657 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6d73fd02-ff75-4664-a80e-a7413dda55d8-audit-policies\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.438114 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-service-ca\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.440140 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-user-template-error\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.441688 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-router-certs\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.442424 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.443094 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-user-template-login\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.443521 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-session\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.444716 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.445882 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.446854 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6d73fd02-ff75-4664-a80e-a7413dda55d8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.466971 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq64n\" (UniqueName: \"kubernetes.io/projected/6d73fd02-ff75-4664-a80e-a7413dda55d8-kube-api-access-sq64n\") pod \"oauth-openshift-57866998d-v97sw\" (UID: \"6d73fd02-ff75-4664-a80e-a7413dda55d8\") " pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.514852 4818 generic.go:334] "Generic (PLEG): container finished" podID="d7f12c1b-f504-4155-bd29-bcb6ab57b646" containerID="8cca0fe7a690cbf136550efe0435f181a102b59974fe93863f2476a130111250" exitCode=0 Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.514904 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" event={"ID":"d7f12c1b-f504-4155-bd29-bcb6ab57b646","Type":"ContainerDied","Data":"8cca0fe7a690cbf136550efe0435f181a102b59974fe93863f2476a130111250"} Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.514931 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" event={"ID":"d7f12c1b-f504-4155-bd29-bcb6ab57b646","Type":"ContainerDied","Data":"4bf926d053399f203eb58cfbe643d78e35f388e3f6483f5cce576a5358ebb2ee"} Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.514949 4818 scope.go:117] "RemoveContainer" containerID="8cca0fe7a690cbf136550efe0435f181a102b59974fe93863f2476a130111250" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.514944 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fv5lc" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.548670 4818 scope.go:117] "RemoveContainer" containerID="8cca0fe7a690cbf136550efe0435f181a102b59974fe93863f2476a130111250" Dec 03 06:31:33 crc kubenswrapper[4818]: E1203 06:31:33.549625 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cca0fe7a690cbf136550efe0435f181a102b59974fe93863f2476a130111250\": container with ID starting with 8cca0fe7a690cbf136550efe0435f181a102b59974fe93863f2476a130111250 not found: ID does not exist" containerID="8cca0fe7a690cbf136550efe0435f181a102b59974fe93863f2476a130111250" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.549673 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cca0fe7a690cbf136550efe0435f181a102b59974fe93863f2476a130111250"} err="failed to get container status \"8cca0fe7a690cbf136550efe0435f181a102b59974fe93863f2476a130111250\": rpc error: code = NotFound desc = could not find container \"8cca0fe7a690cbf136550efe0435f181a102b59974fe93863f2476a130111250\": container with ID starting with 8cca0fe7a690cbf136550efe0435f181a102b59974fe93863f2476a130111250 not found: ID does not exist" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.557755 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fv5lc"] Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.564531 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fv5lc"] Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.570791 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:33 crc kubenswrapper[4818]: I1203 06:31:33.766200 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57866998d-v97sw"] Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.160273 4818 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.160709 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a" gracePeriod=15 Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.160726 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde" gracePeriod=15 Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.160744 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6" gracePeriod=15 Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.160804 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d" gracePeriod=15 Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.160853 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184" gracePeriod=15 Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.162594 4818 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 06:31:34 crc kubenswrapper[4818]: E1203 06:31:34.163040 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.163080 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 06:31:34 crc kubenswrapper[4818]: E1203 06:31:34.163110 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.163129 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 06:31:34 crc kubenswrapper[4818]: E1203 06:31:34.163157 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.163175 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 06:31:34 crc kubenswrapper[4818]: E1203 06:31:34.163200 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.163216 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 06:31:34 crc kubenswrapper[4818]: E1203 06:31:34.163236 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.163253 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 06:31:34 crc kubenswrapper[4818]: E1203 06:31:34.163273 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.163288 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.163519 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.163555 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.163626 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.163644 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.163672 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.163695 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 06:31:34 crc kubenswrapper[4818]: E1203 06:31:34.163987 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.164011 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.168650 4818 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.169456 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.177594 4818 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.247520 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.247590 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.247643 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.247737 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.247783 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.247811 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.247853 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.247888 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.349056 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.349161 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.349208 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.349210 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.349244 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.349316 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.349322 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.349372 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.349351 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.349418 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.349330 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.349496 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.349487 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.349477 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.349556 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.349713 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.529443 4818 generic.go:334] "Generic (PLEG): container finished" podID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" containerID="dd6041b5271c0de55f8a9b33927ef692b055606ec14afcf8cb4a5cb94d4b228f" exitCode=0 Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.529512 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde","Type":"ContainerDied","Data":"dd6041b5271c0de55f8a9b33927ef692b055606ec14afcf8cb4a5cb94d4b228f"} Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.531134 4818 status_manager.go:851] "Failed to get status for pod" podUID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.533006 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.534599 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.535726 4818 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde" exitCode=0 Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.535807 4818 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184" exitCode=0 Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.535864 4818 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6" exitCode=0 Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.535885 4818 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d" exitCode=2 Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.535895 4818 scope.go:117] "RemoveContainer" containerID="27a9c10611f1a7f159a3b14cab4a4ee7607c84c7760a5e886bf47c9de2e30ff5" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.541414 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" event={"ID":"6d73fd02-ff75-4664-a80e-a7413dda55d8","Type":"ContainerStarted","Data":"d1425c0695ae5dc8091979124962541bd4a5334a77a7b55446733c34f191d71b"} Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.541465 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" event={"ID":"6d73fd02-ff75-4664-a80e-a7413dda55d8","Type":"ContainerStarted","Data":"0c8bef8941901a6d35d6d7272757cf60d3be8d7eb16a2bcd68a2b2e586d2b65f"} Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.541730 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.542530 4818 status_manager.go:851] "Failed to get status for pod" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-57866998d-v97sw\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.543105 4818 status_manager.go:851] "Failed to get status for pod" podUID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.564016 4818 patch_prober.go:28] interesting pod/oauth-openshift-57866998d-v97sw container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.57:6443/healthz\": read tcp 10.217.0.2:53692->10.217.0.57:6443: read: connection reset by peer" start-of-body= Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.564074 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.57:6443/healthz\": read tcp 10.217.0.2:53692->10.217.0.57:6443: read: connection reset by peer" Dec 03 06:31:34 crc kubenswrapper[4818]: E1203 06:31:34.564685 4818 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/events\": dial tcp 38.102.83.75:6443: connect: connection refused" event=< Dec 03 06:31:34 crc kubenswrapper[4818]: &Event{ObjectMeta:{oauth-openshift-57866998d-v97sw.187da0dbb92c0838 openshift-authentication 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-authentication,Name:oauth-openshift-57866998d-v97sw,UID:6d73fd02-ff75-4664-a80e-a7413dda55d8,APIVersion:v1,ResourceVersion:29342,FieldPath:spec.containers{oauth-openshift},},Reason:ProbeError,Message:Readiness probe error: Get "https://10.217.0.57:6443/healthz": read tcp 10.217.0.2:53692->10.217.0.57:6443: read: connection reset by peer Dec 03 06:31:34 crc kubenswrapper[4818]: body: Dec 03 06:31:34 crc kubenswrapper[4818]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 06:31:34.564055096 +0000 UTC m=+252.255663858,LastTimestamp:2025-12-03 06:31:34.564055096 +0000 UTC m=+252.255663858,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 03 06:31:34 crc kubenswrapper[4818]: > Dec 03 06:31:34 crc kubenswrapper[4818]: I1203 06:31:34.746375 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7f12c1b-f504-4155-bd29-bcb6ab57b646" path="/var/lib/kubelet/pods/d7f12c1b-f504-4155-bd29-bcb6ab57b646/volumes" Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.551555 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-57866998d-v97sw_6d73fd02-ff75-4664-a80e-a7413dda55d8/oauth-openshift/0.log" Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.551646 4818 generic.go:334] "Generic (PLEG): container finished" podID="6d73fd02-ff75-4664-a80e-a7413dda55d8" containerID="d1425c0695ae5dc8091979124962541bd4a5334a77a7b55446733c34f191d71b" exitCode=255 Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.551762 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" event={"ID":"6d73fd02-ff75-4664-a80e-a7413dda55d8","Type":"ContainerDied","Data":"d1425c0695ae5dc8091979124962541bd4a5334a77a7b55446733c34f191d71b"} Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.552475 4818 scope.go:117] "RemoveContainer" containerID="d1425c0695ae5dc8091979124962541bd4a5334a77a7b55446733c34f191d71b" Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.553089 4818 status_manager.go:851] "Failed to get status for pod" podUID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.553611 4818 status_manager.go:851] "Failed to get status for pod" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-57866998d-v97sw\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.559502 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.833684 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.834436 4818 status_manager.go:851] "Failed to get status for pod" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-57866998d-v97sw\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.835060 4818 status_manager.go:851] "Failed to get status for pod" podUID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.971754 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-var-lock\") pod \"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde\" (UID: \"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde\") " Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.971961 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-var-lock" (OuterVolumeSpecName: "var-lock") pod "24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" (UID: "24a1ab03-bbf6-4d53-9adb-ddc42b90fdde"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.972057 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-kubelet-dir\") pod \"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde\" (UID: \"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde\") " Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.972130 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-kube-api-access\") pod \"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde\" (UID: \"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde\") " Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.972232 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" (UID: "24a1ab03-bbf6-4d53-9adb-ddc42b90fdde"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.972573 4818 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.972611 4818 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:35 crc kubenswrapper[4818]: I1203 06:31:35.983775 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" (UID: "24a1ab03-bbf6-4d53-9adb-ddc42b90fdde"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.073676 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/24a1ab03-bbf6-4d53-9adb-ddc42b90fdde-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.532602 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.533285 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.533960 4818 status_manager.go:851] "Failed to get status for pod" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-57866998d-v97sw\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.534574 4818 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.534873 4818 status_manager.go:851] "Failed to get status for pod" podUID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.568708 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-57866998d-v97sw_6d73fd02-ff75-4664-a80e-a7413dda55d8/oauth-openshift/0.log" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.568800 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" event={"ID":"6d73fd02-ff75-4664-a80e-a7413dda55d8","Type":"ContainerStarted","Data":"b2d51398850dc5e38681576d3da9ef545ba751646d4ce0f6cbb8f1e360e5167e"} Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.569401 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.570120 4818 status_manager.go:851] "Failed to get status for pod" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-57866998d-v97sw\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.570528 4818 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.570756 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"24a1ab03-bbf6-4d53-9adb-ddc42b90fdde","Type":"ContainerDied","Data":"893d0323fd0ce884b024f493e57892b9ee6c04618dc4356e1491f3f0c691a0e8"} Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.570774 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.570787 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="893d0323fd0ce884b024f493e57892b9ee6c04618dc4356e1491f3f0c691a0e8" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.571289 4818 status_manager.go:851] "Failed to get status for pod" podUID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.573036 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.573605 4818 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a" exitCode=0 Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.573655 4818 scope.go:117] "RemoveContainer" containerID="c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.573770 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.585719 4818 status_manager.go:851] "Failed to get status for pod" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-57866998d-v97sw\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.586192 4818 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.586745 4818 status_manager.go:851] "Failed to get status for pod" podUID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.590639 4818 scope.go:117] "RemoveContainer" containerID="83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.606269 4818 scope.go:117] "RemoveContainer" containerID="da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.621196 4818 scope.go:117] "RemoveContainer" containerID="00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.635139 4818 scope.go:117] "RemoveContainer" containerID="64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.660629 4818 scope.go:117] "RemoveContainer" containerID="98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.679924 4818 scope.go:117] "RemoveContainer" containerID="c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde" Dec 03 06:31:36 crc kubenswrapper[4818]: E1203 06:31:36.680418 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\": container with ID starting with c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde not found: ID does not exist" containerID="c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.680449 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde"} err="failed to get container status \"c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\": rpc error: code = NotFound desc = could not find container \"c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde\": container with ID starting with c45f8ce358a6ee4641fe99da175c093685180ae94c2ce7ebb823f2e6bfcb1dde not found: ID does not exist" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.680470 4818 scope.go:117] "RemoveContainer" containerID="83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184" Dec 03 06:31:36 crc kubenswrapper[4818]: E1203 06:31:36.680953 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\": container with ID starting with 83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184 not found: ID does not exist" containerID="83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.680977 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184"} err="failed to get container status \"83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\": rpc error: code = NotFound desc = could not find container \"83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184\": container with ID starting with 83d091f581cb5a406e583c890624fc158cfc4eeb7b894a3951d477bccc63e184 not found: ID does not exist" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.680991 4818 scope.go:117] "RemoveContainer" containerID="da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6" Dec 03 06:31:36 crc kubenswrapper[4818]: E1203 06:31:36.681346 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\": container with ID starting with da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6 not found: ID does not exist" containerID="da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.681370 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6"} err="failed to get container status \"da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\": rpc error: code = NotFound desc = could not find container \"da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6\": container with ID starting with da2c03f854d00c8b3bccb8b12a953802545d43d723a373bd487f0edc92bd27e6 not found: ID does not exist" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.681383 4818 scope.go:117] "RemoveContainer" containerID="00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d" Dec 03 06:31:36 crc kubenswrapper[4818]: E1203 06:31:36.681620 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\": container with ID starting with 00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d not found: ID does not exist" containerID="00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.681640 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d"} err="failed to get container status \"00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\": rpc error: code = NotFound desc = could not find container \"00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d\": container with ID starting with 00e456afef7e9d884d2474e3dc33e914273d3271e2216fd9b66c5b337d9e359d not found: ID does not exist" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.681652 4818 scope.go:117] "RemoveContainer" containerID="64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a" Dec 03 06:31:36 crc kubenswrapper[4818]: E1203 06:31:36.681901 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\": container with ID starting with 64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a not found: ID does not exist" containerID="64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.681924 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a"} err="failed to get container status \"64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\": rpc error: code = NotFound desc = could not find container \"64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a\": container with ID starting with 64aa118c078dc216e648d03d51a3dfa4e81e09c031449603581361509418387a not found: ID does not exist" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.681937 4818 scope.go:117] "RemoveContainer" containerID="98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.682344 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.682367 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.682418 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.682479 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.682594 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.683124 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.683854 4818 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.683959 4818 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.683971 4818 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:31:36 crc kubenswrapper[4818]: E1203 06:31:36.684553 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\": container with ID starting with 98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5 not found: ID does not exist" containerID="98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.684610 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5"} err="failed to get container status \"98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\": rpc error: code = NotFound desc = could not find container \"98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5\": container with ID starting with 98274137e90dd332718a360e49d2bff9bf904a6acc1923f91222470de10e2bf5 not found: ID does not exist" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.746487 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.780009 4818 patch_prober.go:28] interesting pod/oauth-openshift-57866998d-v97sw container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.57:6443/healthz\": read tcp 10.217.0.2:53694->10.217.0.57:6443: read: connection reset by peer" start-of-body= Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.780378 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.57:6443/healthz\": read tcp 10.217.0.2:53694->10.217.0.57:6443: read: connection reset by peer" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.878436 4818 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.879340 4818 status_manager.go:851] "Failed to get status for pod" podUID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:36 crc kubenswrapper[4818]: I1203 06:31:36.879804 4818 status_manager.go:851] "Failed to get status for pod" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-57866998d-v97sw\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:37 crc kubenswrapper[4818]: I1203 06:31:37.579732 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-57866998d-v97sw_6d73fd02-ff75-4664-a80e-a7413dda55d8/oauth-openshift/1.log" Dec 03 06:31:37 crc kubenswrapper[4818]: I1203 06:31:37.580211 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-57866998d-v97sw_6d73fd02-ff75-4664-a80e-a7413dda55d8/oauth-openshift/0.log" Dec 03 06:31:37 crc kubenswrapper[4818]: I1203 06:31:37.580251 4818 generic.go:334] "Generic (PLEG): container finished" podID="6d73fd02-ff75-4664-a80e-a7413dda55d8" containerID="b2d51398850dc5e38681576d3da9ef545ba751646d4ce0f6cbb8f1e360e5167e" exitCode=255 Dec 03 06:31:37 crc kubenswrapper[4818]: I1203 06:31:37.580304 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" event={"ID":"6d73fd02-ff75-4664-a80e-a7413dda55d8","Type":"ContainerDied","Data":"b2d51398850dc5e38681576d3da9ef545ba751646d4ce0f6cbb8f1e360e5167e"} Dec 03 06:31:37 crc kubenswrapper[4818]: I1203 06:31:37.580340 4818 scope.go:117] "RemoveContainer" containerID="d1425c0695ae5dc8091979124962541bd4a5334a77a7b55446733c34f191d71b" Dec 03 06:31:37 crc kubenswrapper[4818]: I1203 06:31:37.580887 4818 scope.go:117] "RemoveContainer" containerID="b2d51398850dc5e38681576d3da9ef545ba751646d4ce0f6cbb8f1e360e5167e" Dec 03 06:31:37 crc kubenswrapper[4818]: I1203 06:31:37.581062 4818 status_manager.go:851] "Failed to get status for pod" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-57866998d-v97sw\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:37 crc kubenswrapper[4818]: E1203 06:31:37.581084 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-57866998d-v97sw_openshift-authentication(6d73fd02-ff75-4664-a80e-a7413dda55d8)\"" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" Dec 03 06:31:37 crc kubenswrapper[4818]: I1203 06:31:37.581332 4818 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:37 crc kubenswrapper[4818]: I1203 06:31:37.581520 4818 status_manager.go:851] "Failed to get status for pod" podUID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:38 crc kubenswrapper[4818]: E1203 06:31:38.433746 4818 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:38 crc kubenswrapper[4818]: E1203 06:31:38.434220 4818 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:38 crc kubenswrapper[4818]: E1203 06:31:38.434580 4818 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:38 crc kubenswrapper[4818]: E1203 06:31:38.434956 4818 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:38 crc kubenswrapper[4818]: E1203 06:31:38.435344 4818 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:38 crc kubenswrapper[4818]: I1203 06:31:38.435403 4818 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 03 06:31:38 crc kubenswrapper[4818]: E1203 06:31:38.435767 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="200ms" Dec 03 06:31:38 crc kubenswrapper[4818]: I1203 06:31:38.593725 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-57866998d-v97sw_6d73fd02-ff75-4664-a80e-a7413dda55d8/oauth-openshift/1.log" Dec 03 06:31:38 crc kubenswrapper[4818]: I1203 06:31:38.594630 4818 scope.go:117] "RemoveContainer" containerID="b2d51398850dc5e38681576d3da9ef545ba751646d4ce0f6cbb8f1e360e5167e" Dec 03 06:31:38 crc kubenswrapper[4818]: E1203 06:31:38.595000 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-57866998d-v97sw_openshift-authentication(6d73fd02-ff75-4664-a80e-a7413dda55d8)\"" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" Dec 03 06:31:38 crc kubenswrapper[4818]: I1203 06:31:38.595158 4818 status_manager.go:851] "Failed to get status for pod" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-57866998d-v97sw\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:38 crc kubenswrapper[4818]: I1203 06:31:38.595733 4818 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:38 crc kubenswrapper[4818]: I1203 06:31:38.596583 4818 status_manager.go:851] "Failed to get status for pod" podUID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:38 crc kubenswrapper[4818]: E1203 06:31:38.636677 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="400ms" Dec 03 06:31:39 crc kubenswrapper[4818]: E1203 06:31:39.037565 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="800ms" Dec 03 06:31:39 crc kubenswrapper[4818]: E1203 06:31:39.199461 4818 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.75:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:39 crc kubenswrapper[4818]: I1203 06:31:39.200189 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:39 crc kubenswrapper[4818]: I1203 06:31:39.602078 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"30444f4bd53c99e184520deb2b39513734f604a8552a737823b529326f183161"} Dec 03 06:31:39 crc kubenswrapper[4818]: I1203 06:31:39.602661 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"2f0eb432247ee1a8bce427b1ec9df824bff1601354f10d18a0376f9c90eb8dcb"} Dec 03 06:31:39 crc kubenswrapper[4818]: I1203 06:31:39.603385 4818 status_manager.go:851] "Failed to get status for pod" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-57866998d-v97sw\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:39 crc kubenswrapper[4818]: E1203 06:31:39.603388 4818 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.75:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:31:39 crc kubenswrapper[4818]: I1203 06:31:39.603715 4818 status_manager.go:851] "Failed to get status for pod" podUID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:39 crc kubenswrapper[4818]: E1203 06:31:39.839105 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="1.6s" Dec 03 06:31:41 crc kubenswrapper[4818]: E1203 06:31:41.440322 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="3.2s" Dec 03 06:31:42 crc kubenswrapper[4818]: I1203 06:31:42.739707 4818 status_manager.go:851] "Failed to get status for pod" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-57866998d-v97sw\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:42 crc kubenswrapper[4818]: I1203 06:31:42.740162 4818 status_manager.go:851] "Failed to get status for pod" podUID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:43 crc kubenswrapper[4818]: I1203 06:31:43.571302 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:43 crc kubenswrapper[4818]: I1203 06:31:43.572753 4818 scope.go:117] "RemoveContainer" containerID="b2d51398850dc5e38681576d3da9ef545ba751646d4ce0f6cbb8f1e360e5167e" Dec 03 06:31:43 crc kubenswrapper[4818]: E1203 06:31:43.573292 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-57866998d-v97sw_openshift-authentication(6d73fd02-ff75-4664-a80e-a7413dda55d8)\"" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" Dec 03 06:31:44 crc kubenswrapper[4818]: E1203 06:31:44.120197 4818 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/events\": dial tcp 38.102.83.75:6443: connect: connection refused" event=< Dec 03 06:31:44 crc kubenswrapper[4818]: &Event{ObjectMeta:{oauth-openshift-57866998d-v97sw.187da0dbb92c0838 openshift-authentication 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-authentication,Name:oauth-openshift-57866998d-v97sw,UID:6d73fd02-ff75-4664-a80e-a7413dda55d8,APIVersion:v1,ResourceVersion:29342,FieldPath:spec.containers{oauth-openshift},},Reason:ProbeError,Message:Readiness probe error: Get "https://10.217.0.57:6443/healthz": read tcp 10.217.0.2:53692->10.217.0.57:6443: read: connection reset by peer Dec 03 06:31:44 crc kubenswrapper[4818]: body: Dec 03 06:31:44 crc kubenswrapper[4818]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 06:31:34.564055096 +0000 UTC m=+252.255663858,LastTimestamp:2025-12-03 06:31:34.564055096 +0000 UTC m=+252.255663858,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 03 06:31:44 crc kubenswrapper[4818]: > Dec 03 06:31:44 crc kubenswrapper[4818]: E1203 06:31:44.642770 4818 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.75:6443: connect: connection refused" interval="6.4s" Dec 03 06:31:46 crc kubenswrapper[4818]: I1203 06:31:46.737335 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:46 crc kubenswrapper[4818]: I1203 06:31:46.738778 4818 status_manager.go:851] "Failed to get status for pod" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-57866998d-v97sw\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:46 crc kubenswrapper[4818]: I1203 06:31:46.739529 4818 status_manager.go:851] "Failed to get status for pod" podUID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:46 crc kubenswrapper[4818]: I1203 06:31:46.763231 4818 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="558ef6b2-cb0b-4631-b025-8fe23b244a88" Dec 03 06:31:46 crc kubenswrapper[4818]: I1203 06:31:46.763279 4818 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="558ef6b2-cb0b-4631-b025-8fe23b244a88" Dec 03 06:31:46 crc kubenswrapper[4818]: E1203 06:31:46.763699 4818 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:46 crc kubenswrapper[4818]: I1203 06:31:46.764450 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:46 crc kubenswrapper[4818]: W1203 06:31:46.800124 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-1537a20d93fb9bc512dcb414c58dec3b7b8164705dbcf1d1014613d351858e66 WatchSource:0}: Error finding container 1537a20d93fb9bc512dcb414c58dec3b7b8164705dbcf1d1014613d351858e66: Status 404 returned error can't find the container with id 1537a20d93fb9bc512dcb414c58dec3b7b8164705dbcf1d1014613d351858e66 Dec 03 06:31:47 crc kubenswrapper[4818]: I1203 06:31:47.662682 4818 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="9fef79dde74afafc3a57028f925e4878ac1564fda26afb232430d015731ab6f1" exitCode=0 Dec 03 06:31:47 crc kubenswrapper[4818]: I1203 06:31:47.662810 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"9fef79dde74afafc3a57028f925e4878ac1564fda26afb232430d015731ab6f1"} Dec 03 06:31:47 crc kubenswrapper[4818]: I1203 06:31:47.663116 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1537a20d93fb9bc512dcb414c58dec3b7b8164705dbcf1d1014613d351858e66"} Dec 03 06:31:47 crc kubenswrapper[4818]: I1203 06:31:47.663542 4818 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="558ef6b2-cb0b-4631-b025-8fe23b244a88" Dec 03 06:31:47 crc kubenswrapper[4818]: I1203 06:31:47.663568 4818 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="558ef6b2-cb0b-4631-b025-8fe23b244a88" Dec 03 06:31:47 crc kubenswrapper[4818]: I1203 06:31:47.664154 4818 status_manager.go:851] "Failed to get status for pod" podUID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:47 crc kubenswrapper[4818]: E1203 06:31:47.664507 4818 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.75:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:47 crc kubenswrapper[4818]: I1203 06:31:47.664698 4818 status_manager.go:851] "Failed to get status for pod" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-57866998d-v97sw\": dial tcp 38.102.83.75:6443: connect: connection refused" Dec 03 06:31:48 crc kubenswrapper[4818]: I1203 06:31:48.671615 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"52b03cacb25cddf8d603da2177b23d5b3d4a41660eee91b1cb0cc47c794ad5ad"} Dec 03 06:31:48 crc kubenswrapper[4818]: I1203 06:31:48.672171 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"56ee7ffd1a99baddfe5784b97190976541f08085e1bf16e5f7355c99baa1d3ea"} Dec 03 06:31:48 crc kubenswrapper[4818]: I1203 06:31:48.672184 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"aa03b15752c76dd15857b6ee4ed992c160570e6f0178bafbcf6f74769d9f13a0"} Dec 03 06:31:48 crc kubenswrapper[4818]: I1203 06:31:48.672197 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"642ab5d6d52de3227dcdfc8ec84e11c687af5800973d81f8b4edb199d1b5e87a"} Dec 03 06:31:49 crc kubenswrapper[4818]: I1203 06:31:49.680382 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"873be4ad5027920217e52ce9eab787996af6031ad9b92a89fa9e2c111fd3d47a"} Dec 03 06:31:49 crc kubenswrapper[4818]: I1203 06:31:49.680545 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:49 crc kubenswrapper[4818]: I1203 06:31:49.680672 4818 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="558ef6b2-cb0b-4631-b025-8fe23b244a88" Dec 03 06:31:49 crc kubenswrapper[4818]: I1203 06:31:49.680696 4818 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="558ef6b2-cb0b-4631-b025-8fe23b244a88" Dec 03 06:31:49 crc kubenswrapper[4818]: I1203 06:31:49.683741 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 06:31:49 crc kubenswrapper[4818]: I1203 06:31:49.683791 4818 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86" exitCode=1 Dec 03 06:31:49 crc kubenswrapper[4818]: I1203 06:31:49.683868 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86"} Dec 03 06:31:49 crc kubenswrapper[4818]: I1203 06:31:49.684634 4818 scope.go:117] "RemoveContainer" containerID="7133e17d2dcb0a5a44c260b55eb3824807186767ce60fcebfdd6f9ed56c7cb86" Dec 03 06:31:50 crc kubenswrapper[4818]: I1203 06:31:50.688219 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:31:50 crc kubenswrapper[4818]: I1203 06:31:50.688593 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:31:50 crc kubenswrapper[4818]: I1203 06:31:50.688626 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:31:50 crc kubenswrapper[4818]: I1203 06:31:50.688667 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:31:50 crc kubenswrapper[4818]: I1203 06:31:50.693485 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 06:31:50 crc kubenswrapper[4818]: I1203 06:31:50.693549 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e11fb04b2e3d1979a292a00dbb0ec8d64e307abd1273c9be6f2c0318004904e3"} Dec 03 06:31:51 crc kubenswrapper[4818]: E1203 06:31:51.689523 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 03 06:31:51 crc kubenswrapper[4818]: E1203 06:31:51.689581 4818 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 03 06:31:51 crc kubenswrapper[4818]: E1203 06:31:51.689646 4818 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: failed to sync configmap cache: timed out waiting for the condition Dec 03 06:31:51 crc kubenswrapper[4818]: E1203 06:31:51.689702 4818 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: failed to sync secret cache: timed out waiting for the condition Dec 03 06:31:51 crc kubenswrapper[4818]: E1203 06:31:51.690406 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:33:53.6903277 +0000 UTC m=+391.381936492 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : failed to sync configmap cache: timed out waiting for the condition Dec 03 06:31:51 crc kubenswrapper[4818]: E1203 06:31:51.690460 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 06:33:53.690424112 +0000 UTC m=+391.382032894 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : failed to sync secret cache: timed out waiting for the condition Dec 03 06:31:51 crc kubenswrapper[4818]: I1203 06:31:51.694107 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 06:31:51 crc kubenswrapper[4818]: E1203 06:31:51.700250 4818 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: failed to sync configmap cache: timed out waiting for the condition Dec 03 06:31:51 crc kubenswrapper[4818]: E1203 06:31:51.700251 4818 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: failed to sync configmap cache: timed out waiting for the condition Dec 03 06:31:51 crc kubenswrapper[4818]: E1203 06:31:51.700347 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 06:33:53.700323378 +0000 UTC m=+391.391932130 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : failed to sync configmap cache: timed out waiting for the condition Dec 03 06:31:51 crc kubenswrapper[4818]: E1203 06:31:51.700407 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 06:33:53.700380969 +0000 UTC m=+391.391989951 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : failed to sync configmap cache: timed out waiting for the condition Dec 03 06:31:51 crc kubenswrapper[4818]: E1203 06:31:51.757204 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-s2dwl], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 06:31:51 crc kubenswrapper[4818]: I1203 06:31:51.764999 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:51 crc kubenswrapper[4818]: I1203 06:31:51.765050 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:51 crc kubenswrapper[4818]: I1203 06:31:51.773348 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:52 crc kubenswrapper[4818]: E1203 06:31:52.774249 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-cqllr], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 06:31:52 crc kubenswrapper[4818]: E1203 06:31:52.787190 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[networking-console-plugin-cert nginx-conf], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 06:31:54 crc kubenswrapper[4818]: I1203 06:31:54.698693 4818 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:54 crc kubenswrapper[4818]: I1203 06:31:54.754913 4818 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="cdb805bb-1743-462f-8cd4-0202a76b58ea" Dec 03 06:31:55 crc kubenswrapper[4818]: I1203 06:31:55.031751 4818 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="558ef6b2-cb0b-4631-b025-8fe23b244a88" Dec 03 06:31:55 crc kubenswrapper[4818]: I1203 06:31:55.031801 4818 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="558ef6b2-cb0b-4631-b025-8fe23b244a88" Dec 03 06:31:55 crc kubenswrapper[4818]: I1203 06:31:55.035617 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:31:55 crc kubenswrapper[4818]: I1203 06:31:55.035735 4818 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="cdb805bb-1743-462f-8cd4-0202a76b58ea" Dec 03 06:31:55 crc kubenswrapper[4818]: I1203 06:31:55.227703 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:31:55 crc kubenswrapper[4818]: I1203 06:31:55.234622 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:31:55 crc kubenswrapper[4818]: I1203 06:31:55.693151 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 06:31:55 crc kubenswrapper[4818]: I1203 06:31:55.693327 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 06:31:55 crc kubenswrapper[4818]: I1203 06:31:55.693709 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 06:31:56 crc kubenswrapper[4818]: I1203 06:31:56.036679 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:31:56 crc kubenswrapper[4818]: I1203 06:31:56.036950 4818 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="558ef6b2-cb0b-4631-b025-8fe23b244a88" Dec 03 06:31:56 crc kubenswrapper[4818]: I1203 06:31:56.036974 4818 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="558ef6b2-cb0b-4631-b025-8fe23b244a88" Dec 03 06:31:56 crc kubenswrapper[4818]: I1203 06:31:56.039509 4818 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="cdb805bb-1743-462f-8cd4-0202a76b58ea" Dec 03 06:31:57 crc kubenswrapper[4818]: I1203 06:31:57.737702 4818 scope.go:117] "RemoveContainer" containerID="b2d51398850dc5e38681576d3da9ef545ba751646d4ce0f6cbb8f1e360e5167e" Dec 03 06:31:58 crc kubenswrapper[4818]: I1203 06:31:58.056132 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-57866998d-v97sw_6d73fd02-ff75-4664-a80e-a7413dda55d8/oauth-openshift/1.log" Dec 03 06:31:58 crc kubenswrapper[4818]: I1203 06:31:58.056724 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" event={"ID":"6d73fd02-ff75-4664-a80e-a7413dda55d8","Type":"ContainerStarted","Data":"cecff1669eef6cdffbab025622f48f25e6104f558cfef53b42b88f68764c137e"} Dec 03 06:31:58 crc kubenswrapper[4818]: I1203 06:31:58.057732 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:31:58 crc kubenswrapper[4818]: I1203 06:31:58.058413 4818 patch_prober.go:28] interesting pod/oauth-openshift-57866998d-v97sw container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.57:6443/healthz\": dial tcp 10.217.0.57:6443: connect: connection refused" start-of-body= Dec 03 06:31:58 crc kubenswrapper[4818]: I1203 06:31:58.058478 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.57:6443/healthz\": dial tcp 10.217.0.57:6443: connect: connection refused" Dec 03 06:31:59 crc kubenswrapper[4818]: I1203 06:31:59.067293 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-57866998d-v97sw_6d73fd02-ff75-4664-a80e-a7413dda55d8/oauth-openshift/2.log" Dec 03 06:31:59 crc kubenswrapper[4818]: I1203 06:31:59.068484 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-57866998d-v97sw_6d73fd02-ff75-4664-a80e-a7413dda55d8/oauth-openshift/1.log" Dec 03 06:31:59 crc kubenswrapper[4818]: I1203 06:31:59.068572 4818 generic.go:334] "Generic (PLEG): container finished" podID="6d73fd02-ff75-4664-a80e-a7413dda55d8" containerID="cecff1669eef6cdffbab025622f48f25e6104f558cfef53b42b88f68764c137e" exitCode=255 Dec 03 06:31:59 crc kubenswrapper[4818]: I1203 06:31:59.068622 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" event={"ID":"6d73fd02-ff75-4664-a80e-a7413dda55d8","Type":"ContainerDied","Data":"cecff1669eef6cdffbab025622f48f25e6104f558cfef53b42b88f68764c137e"} Dec 03 06:31:59 crc kubenswrapper[4818]: I1203 06:31:59.068680 4818 scope.go:117] "RemoveContainer" containerID="b2d51398850dc5e38681576d3da9ef545ba751646d4ce0f6cbb8f1e360e5167e" Dec 03 06:31:59 crc kubenswrapper[4818]: I1203 06:31:59.069421 4818 scope.go:117] "RemoveContainer" containerID="cecff1669eef6cdffbab025622f48f25e6104f558cfef53b42b88f68764c137e" Dec 03 06:31:59 crc kubenswrapper[4818]: E1203 06:31:59.069859 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 20s restarting failed container=oauth-openshift pod=oauth-openshift-57866998d-v97sw_openshift-authentication(6d73fd02-ff75-4664-a80e-a7413dda55d8)\"" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" Dec 03 06:32:00 crc kubenswrapper[4818]: I1203 06:32:00.075580 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-57866998d-v97sw_6d73fd02-ff75-4664-a80e-a7413dda55d8/oauth-openshift/2.log" Dec 03 06:32:00 crc kubenswrapper[4818]: I1203 06:32:00.076802 4818 scope.go:117] "RemoveContainer" containerID="cecff1669eef6cdffbab025622f48f25e6104f558cfef53b42b88f68764c137e" Dec 03 06:32:00 crc kubenswrapper[4818]: E1203 06:32:00.077074 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 20s restarting failed container=oauth-openshift pod=oauth-openshift-57866998d-v97sw_openshift-authentication(6d73fd02-ff75-4664-a80e-a7413dda55d8)\"" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" Dec 03 06:32:02 crc kubenswrapper[4818]: I1203 06:32:02.739260 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:32:03 crc kubenswrapper[4818]: I1203 06:32:03.571453 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:32:03 crc kubenswrapper[4818]: I1203 06:32:03.572253 4818 scope.go:117] "RemoveContainer" containerID="cecff1669eef6cdffbab025622f48f25e6104f558cfef53b42b88f68764c137e" Dec 03 06:32:03 crc kubenswrapper[4818]: E1203 06:32:03.572556 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 20s restarting failed container=oauth-openshift pod=oauth-openshift-57866998d-v97sw_openshift-authentication(6d73fd02-ff75-4664-a80e-a7413dda55d8)\"" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" Dec 03 06:32:03 crc kubenswrapper[4818]: I1203 06:32:03.737501 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:32:04 crc kubenswrapper[4818]: I1203 06:32:04.183652 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 06:32:04 crc kubenswrapper[4818]: I1203 06:32:04.611844 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 06:32:04 crc kubenswrapper[4818]: I1203 06:32:04.696967 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 06:32:04 crc kubenswrapper[4818]: I1203 06:32:04.737997 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:32:05 crc kubenswrapper[4818]: I1203 06:32:05.165905 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 06:32:05 crc kubenswrapper[4818]: I1203 06:32:05.212524 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 06:32:05 crc kubenswrapper[4818]: I1203 06:32:05.356097 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 06:32:05 crc kubenswrapper[4818]: I1203 06:32:05.425456 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 06:32:05 crc kubenswrapper[4818]: I1203 06:32:05.498760 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 06:32:05 crc kubenswrapper[4818]: I1203 06:32:05.844809 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 06:32:05 crc kubenswrapper[4818]: I1203 06:32:05.966159 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 06:32:06 crc kubenswrapper[4818]: I1203 06:32:06.145430 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 06:32:06 crc kubenswrapper[4818]: I1203 06:32:06.187536 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 06:32:06 crc kubenswrapper[4818]: I1203 06:32:06.250958 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 06:32:06 crc kubenswrapper[4818]: I1203 06:32:06.489516 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 06:32:06 crc kubenswrapper[4818]: I1203 06:32:06.654690 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 06:32:06 crc kubenswrapper[4818]: I1203 06:32:06.854514 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 06:32:06 crc kubenswrapper[4818]: I1203 06:32:06.885631 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.066978 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.096510 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.122749 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.236023 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.370048 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.427173 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.507788 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.547807 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.555268 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.581725 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.614074 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.724437 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.730012 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.831194 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.844900 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.855643 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.951520 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.975156 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 06:32:07 crc kubenswrapper[4818]: I1203 06:32:07.986455 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 06:32:08 crc kubenswrapper[4818]: I1203 06:32:08.005904 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 06:32:08 crc kubenswrapper[4818]: I1203 06:32:08.142657 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 06:32:08 crc kubenswrapper[4818]: I1203 06:32:08.347397 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 06:32:08 crc kubenswrapper[4818]: I1203 06:32:08.508715 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 06:32:08 crc kubenswrapper[4818]: I1203 06:32:08.524455 4818 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 06:32:08 crc kubenswrapper[4818]: I1203 06:32:08.525965 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 06:32:08 crc kubenswrapper[4818]: I1203 06:32:08.533774 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 06:32:08 crc kubenswrapper[4818]: I1203 06:32:08.610151 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 06:32:08 crc kubenswrapper[4818]: I1203 06:32:08.721371 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 06:32:08 crc kubenswrapper[4818]: I1203 06:32:08.808236 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 06:32:08 crc kubenswrapper[4818]: I1203 06:32:08.819420 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 06:32:08 crc kubenswrapper[4818]: I1203 06:32:08.908052 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 06:32:08 crc kubenswrapper[4818]: I1203 06:32:08.973096 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 06:32:08 crc kubenswrapper[4818]: I1203 06:32:08.976604 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 06:32:08 crc kubenswrapper[4818]: I1203 06:32:08.988746 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 06:32:09 crc kubenswrapper[4818]: I1203 06:32:09.045858 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 06:32:09 crc kubenswrapper[4818]: I1203 06:32:09.208358 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 06:32:09 crc kubenswrapper[4818]: I1203 06:32:09.251667 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 06:32:09 crc kubenswrapper[4818]: I1203 06:32:09.293706 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 06:32:09 crc kubenswrapper[4818]: I1203 06:32:09.330232 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 06:32:09 crc kubenswrapper[4818]: I1203 06:32:09.403721 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 06:32:09 crc kubenswrapper[4818]: I1203 06:32:09.436320 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 06:32:09 crc kubenswrapper[4818]: I1203 06:32:09.811340 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 06:32:09 crc kubenswrapper[4818]: I1203 06:32:09.959482 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.164881 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.195866 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.227912 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.232196 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.235778 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.249874 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.262761 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.293859 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.340663 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.341968 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.374674 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.413893 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.447860 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.739275 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.745536 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.831583 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.834352 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.918769 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.940406 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.958155 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 06:32:10 crc kubenswrapper[4818]: I1203 06:32:10.967859 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.047534 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.066516 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.070361 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.163696 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.200527 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.231770 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.232909 4818 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.279568 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.322898 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.402080 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.428594 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.471908 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.557297 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.584028 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.602375 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.662573 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.825720 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.888216 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.903957 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.904316 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 06:32:11 crc kubenswrapper[4818]: I1203 06:32:11.970959 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.032171 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.043018 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.057535 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.187131 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.269687 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.295279 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.299638 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.342693 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.396938 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.469043 4818 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.473593 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.473647 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.478108 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.478295 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.516351 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=18.516329206 podStartE2EDuration="18.516329206s" podCreationTimestamp="2025-12-03 06:31:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:32:12.494100296 +0000 UTC m=+290.185709058" watchObservedRunningTime="2025-12-03 06:32:12.516329206 +0000 UTC m=+290.207937958" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.526974 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.597271 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.636326 4818 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.642943 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.656595 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.690850 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.808991 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.829756 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 06:32:12 crc kubenswrapper[4818]: I1203 06:32:12.922360 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 06:32:13 crc kubenswrapper[4818]: I1203 06:32:13.079959 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 06:32:13 crc kubenswrapper[4818]: I1203 06:32:13.122833 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 06:32:13 crc kubenswrapper[4818]: I1203 06:32:13.202316 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 06:32:13 crc kubenswrapper[4818]: I1203 06:32:13.296893 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 06:32:13 crc kubenswrapper[4818]: I1203 06:32:13.317922 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 06:32:13 crc kubenswrapper[4818]: I1203 06:32:13.350780 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 06:32:13 crc kubenswrapper[4818]: I1203 06:32:13.601347 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 06:32:13 crc kubenswrapper[4818]: I1203 06:32:13.747460 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 06:32:13 crc kubenswrapper[4818]: I1203 06:32:13.752356 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 06:32:13 crc kubenswrapper[4818]: I1203 06:32:13.771372 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 06:32:13 crc kubenswrapper[4818]: I1203 06:32:13.847066 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 06:32:13 crc kubenswrapper[4818]: I1203 06:32:13.922620 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 06:32:13 crc kubenswrapper[4818]: I1203 06:32:13.924165 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 06:32:14 crc kubenswrapper[4818]: I1203 06:32:14.016228 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 06:32:14 crc kubenswrapper[4818]: I1203 06:32:14.022117 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 06:32:14 crc kubenswrapper[4818]: I1203 06:32:14.112053 4818 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 06:32:14 crc kubenswrapper[4818]: I1203 06:32:14.226804 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 06:32:14 crc kubenswrapper[4818]: I1203 06:32:14.232017 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 06:32:14 crc kubenswrapper[4818]: I1203 06:32:14.336090 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 06:32:14 crc kubenswrapper[4818]: I1203 06:32:14.607563 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 06:32:14 crc kubenswrapper[4818]: I1203 06:32:14.634563 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 06:32:14 crc kubenswrapper[4818]: I1203 06:32:14.657176 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 06:32:14 crc kubenswrapper[4818]: I1203 06:32:14.686151 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 06:32:14 crc kubenswrapper[4818]: I1203 06:32:14.827386 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 06:32:14 crc kubenswrapper[4818]: I1203 06:32:14.896377 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 06:32:14 crc kubenswrapper[4818]: I1203 06:32:14.904608 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 06:32:14 crc kubenswrapper[4818]: I1203 06:32:14.941567 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 06:32:15 crc kubenswrapper[4818]: I1203 06:32:15.029459 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 06:32:15 crc kubenswrapper[4818]: I1203 06:32:15.046358 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 06:32:15 crc kubenswrapper[4818]: I1203 06:32:15.318424 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 06:32:15 crc kubenswrapper[4818]: I1203 06:32:15.318536 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 06:32:15 crc kubenswrapper[4818]: I1203 06:32:15.318577 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 06:32:15 crc kubenswrapper[4818]: I1203 06:32:15.340914 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 06:32:15 crc kubenswrapper[4818]: I1203 06:32:15.436053 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 06:32:15 crc kubenswrapper[4818]: I1203 06:32:15.445265 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 06:32:15 crc kubenswrapper[4818]: I1203 06:32:15.517658 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 06:32:15 crc kubenswrapper[4818]: I1203 06:32:15.558259 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 06:32:15 crc kubenswrapper[4818]: I1203 06:32:15.580388 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 06:32:15 crc kubenswrapper[4818]: I1203 06:32:15.727206 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 06:32:15 crc kubenswrapper[4818]: I1203 06:32:15.727610 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 06:32:15 crc kubenswrapper[4818]: I1203 06:32:15.734304 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 06:32:15 crc kubenswrapper[4818]: I1203 06:32:15.791225 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 06:32:15 crc kubenswrapper[4818]: I1203 06:32:15.930766 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.002146 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.051013 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.215495 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.225132 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.262976 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.265507 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.267742 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.354338 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.377299 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.396555 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.410774 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.448424 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.449951 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.487015 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.507739 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.604491 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.654941 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.695792 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.737416 4818 scope.go:117] "RemoveContainer" containerID="cecff1669eef6cdffbab025622f48f25e6104f558cfef53b42b88f68764c137e" Dec 03 06:32:16 crc kubenswrapper[4818]: E1203 06:32:16.737953 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 20s restarting failed container=oauth-openshift pod=oauth-openshift-57866998d-v97sw_openshift-authentication(6d73fd02-ff75-4664-a80e-a7413dda55d8)\"" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" podUID="6d73fd02-ff75-4664-a80e-a7413dda55d8" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.786179 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.810409 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.819545 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.916401 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 06:32:16 crc kubenswrapper[4818]: I1203 06:32:16.987102 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.022604 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.058901 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.071647 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.110624 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.111634 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.222346 4818 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.222709 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://30444f4bd53c99e184520deb2b39513734f604a8552a737823b529326f183161" gracePeriod=5 Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.237100 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.284719 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.310791 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.351348 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.351496 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.502761 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.551270 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.558987 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.559765 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.592735 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.643806 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.679249 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.685026 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.687546 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.696673 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.811163 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.854578 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.959770 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 06:32:17 crc kubenswrapper[4818]: I1203 06:32:17.979328 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 06:32:18 crc kubenswrapper[4818]: I1203 06:32:18.027061 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 06:32:18 crc kubenswrapper[4818]: I1203 06:32:18.028654 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 06:32:18 crc kubenswrapper[4818]: I1203 06:32:18.113179 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 06:32:18 crc kubenswrapper[4818]: I1203 06:32:18.267266 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 06:32:18 crc kubenswrapper[4818]: I1203 06:32:18.268214 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 06:32:18 crc kubenswrapper[4818]: I1203 06:32:18.348254 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 06:32:18 crc kubenswrapper[4818]: I1203 06:32:18.453707 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 06:32:18 crc kubenswrapper[4818]: I1203 06:32:18.461521 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 06:32:18 crc kubenswrapper[4818]: I1203 06:32:18.531091 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 06:32:18 crc kubenswrapper[4818]: I1203 06:32:18.566999 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 06:32:18 crc kubenswrapper[4818]: I1203 06:32:18.761043 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 06:32:18 crc kubenswrapper[4818]: I1203 06:32:18.796734 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 06:32:18 crc kubenswrapper[4818]: I1203 06:32:18.841085 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 06:32:18 crc kubenswrapper[4818]: I1203 06:32:18.889917 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 06:32:19 crc kubenswrapper[4818]: I1203 06:32:19.020134 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 06:32:19 crc kubenswrapper[4818]: I1203 06:32:19.167729 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 06:32:19 crc kubenswrapper[4818]: I1203 06:32:19.191497 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 06:32:19 crc kubenswrapper[4818]: I1203 06:32:19.463511 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 06:32:19 crc kubenswrapper[4818]: I1203 06:32:19.476646 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 06:32:19 crc kubenswrapper[4818]: I1203 06:32:19.599764 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 06:32:19 crc kubenswrapper[4818]: I1203 06:32:19.610207 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 06:32:19 crc kubenswrapper[4818]: I1203 06:32:19.624080 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 06:32:19 crc kubenswrapper[4818]: I1203 06:32:19.660339 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 06:32:19 crc kubenswrapper[4818]: I1203 06:32:19.711455 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 06:32:20 crc kubenswrapper[4818]: I1203 06:32:20.270566 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 06:32:20 crc kubenswrapper[4818]: I1203 06:32:20.284694 4818 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 06:32:20 crc kubenswrapper[4818]: I1203 06:32:20.344317 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 06:32:20 crc kubenswrapper[4818]: I1203 06:32:20.434370 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 06:32:20 crc kubenswrapper[4818]: I1203 06:32:20.518189 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 06:32:20 crc kubenswrapper[4818]: I1203 06:32:20.571959 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 06:32:20 crc kubenswrapper[4818]: I1203 06:32:20.670224 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 06:32:20 crc kubenswrapper[4818]: I1203 06:32:20.733232 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 06:32:20 crc kubenswrapper[4818]: I1203 06:32:20.759238 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 06:32:21 crc kubenswrapper[4818]: I1203 06:32:21.437700 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 06:32:21 crc kubenswrapper[4818]: I1203 06:32:21.960595 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.146284 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.388025 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.816881 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.816946 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.826264 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.826312 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.826327 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.826351 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.826375 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.826492 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.826531 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.826521 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.826717 4818 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.826741 4818 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.826758 4818 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.826788 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.834086 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.927973 4818 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 06:32:22 crc kubenswrapper[4818]: I1203 06:32:22.928289 4818 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 03 06:32:23 crc kubenswrapper[4818]: I1203 06:32:23.205627 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 06:32:23 crc kubenswrapper[4818]: I1203 06:32:23.205680 4818 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="30444f4bd53c99e184520deb2b39513734f604a8552a737823b529326f183161" exitCode=137 Dec 03 06:32:23 crc kubenswrapper[4818]: I1203 06:32:23.205722 4818 scope.go:117] "RemoveContainer" containerID="30444f4bd53c99e184520deb2b39513734f604a8552a737823b529326f183161" Dec 03 06:32:23 crc kubenswrapper[4818]: I1203 06:32:23.205842 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 06:32:23 crc kubenswrapper[4818]: I1203 06:32:23.232288 4818 scope.go:117] "RemoveContainer" containerID="30444f4bd53c99e184520deb2b39513734f604a8552a737823b529326f183161" Dec 03 06:32:23 crc kubenswrapper[4818]: E1203 06:32:23.232686 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30444f4bd53c99e184520deb2b39513734f604a8552a737823b529326f183161\": container with ID starting with 30444f4bd53c99e184520deb2b39513734f604a8552a737823b529326f183161 not found: ID does not exist" containerID="30444f4bd53c99e184520deb2b39513734f604a8552a737823b529326f183161" Dec 03 06:32:23 crc kubenswrapper[4818]: I1203 06:32:23.232723 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30444f4bd53c99e184520deb2b39513734f604a8552a737823b529326f183161"} err="failed to get container status \"30444f4bd53c99e184520deb2b39513734f604a8552a737823b529326f183161\": rpc error: code = NotFound desc = could not find container \"30444f4bd53c99e184520deb2b39513734f604a8552a737823b529326f183161\": container with ID starting with 30444f4bd53c99e184520deb2b39513734f604a8552a737823b529326f183161 not found: ID does not exist" Dec 03 06:32:24 crc kubenswrapper[4818]: I1203 06:32:24.745630 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 03 06:32:27 crc kubenswrapper[4818]: I1203 06:32:27.737653 4818 scope.go:117] "RemoveContainer" containerID="cecff1669eef6cdffbab025622f48f25e6104f558cfef53b42b88f68764c137e" Dec 03 06:32:28 crc kubenswrapper[4818]: I1203 06:32:28.240134 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-57866998d-v97sw_6d73fd02-ff75-4664-a80e-a7413dda55d8/oauth-openshift/2.log" Dec 03 06:32:28 crc kubenswrapper[4818]: I1203 06:32:28.240420 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" event={"ID":"6d73fd02-ff75-4664-a80e-a7413dda55d8","Type":"ContainerStarted","Data":"7e0c4623757dcdb3c20ce2d1e9b2085bfab9865b65ac15fc29a807d344c65a91"} Dec 03 06:32:28 crc kubenswrapper[4818]: I1203 06:32:28.241380 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:32:28 crc kubenswrapper[4818]: I1203 06:32:28.266977 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" podStartSLOduration=81.266944866 podStartE2EDuration="1m21.266944866s" podCreationTimestamp="2025-12-03 06:31:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:31:58.080717614 +0000 UTC m=+275.772326366" watchObservedRunningTime="2025-12-03 06:32:28.266944866 +0000 UTC m=+305.958553708" Dec 03 06:32:28 crc kubenswrapper[4818]: I1203 06:32:28.612167 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-57866998d-v97sw" Dec 03 06:32:37 crc kubenswrapper[4818]: I1203 06:32:37.295127 4818 generic.go:334] "Generic (PLEG): container finished" podID="53a413a7-34f8-4fb3-893b-c4fa9711017a" containerID="dddd0b228babbc5d22cd3ceeca18a47a59bd31fc0b34de7805a71b039c0c2fcb" exitCode=0 Dec 03 06:32:37 crc kubenswrapper[4818]: I1203 06:32:37.295678 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" event={"ID":"53a413a7-34f8-4fb3-893b-c4fa9711017a","Type":"ContainerDied","Data":"dddd0b228babbc5d22cd3ceeca18a47a59bd31fc0b34de7805a71b039c0c2fcb"} Dec 03 06:32:37 crc kubenswrapper[4818]: I1203 06:32:37.296171 4818 scope.go:117] "RemoveContainer" containerID="dddd0b228babbc5d22cd3ceeca18a47a59bd31fc0b34de7805a71b039c0c2fcb" Dec 03 06:32:38 crc kubenswrapper[4818]: I1203 06:32:38.305790 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" event={"ID":"53a413a7-34f8-4fb3-893b-c4fa9711017a","Type":"ContainerStarted","Data":"3395f8e86e03b477f14f091b8e9c92d7de78b57f50e6601bb62d2c7e1dbad671"} Dec 03 06:32:38 crc kubenswrapper[4818]: I1203 06:32:38.306809 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" Dec 03 06:32:38 crc kubenswrapper[4818]: I1203 06:32:38.310510 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.173578 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cwhmk"] Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.255286 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58"] Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.255513 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" podUID="6a406cec-8381-4a82-8d46-3ababcda72b4" containerName="route-controller-manager" containerID="cri-o://8941867e59b8f0abdf51e22f253e435eea713ca60abc759608fac0bbc364f02a" gracePeriod=30 Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.426200 4818 generic.go:334] "Generic (PLEG): container finished" podID="6a406cec-8381-4a82-8d46-3ababcda72b4" containerID="8941867e59b8f0abdf51e22f253e435eea713ca60abc759608fac0bbc364f02a" exitCode=0 Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.426280 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" event={"ID":"6a406cec-8381-4a82-8d46-3ababcda72b4","Type":"ContainerDied","Data":"8941867e59b8f0abdf51e22f253e435eea713ca60abc759608fac0bbc364f02a"} Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.426410 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" podUID="a8fc6bf6-a48e-4f93-8a6c-57614d1171b7" containerName="controller-manager" containerID="cri-o://37f806687d01927c0c9fe9c57dd2e12b9ce2f0a32b8287d06ea0ca1993b5d98b" gracePeriod=30 Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.605145 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.730839 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.752087 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a406cec-8381-4a82-8d46-3ababcda72b4-serving-cert\") pod \"6a406cec-8381-4a82-8d46-3ababcda72b4\" (UID: \"6a406cec-8381-4a82-8d46-3ababcda72b4\") " Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.752142 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a406cec-8381-4a82-8d46-3ababcda72b4-client-ca\") pod \"6a406cec-8381-4a82-8d46-3ababcda72b4\" (UID: \"6a406cec-8381-4a82-8d46-3ababcda72b4\") " Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.752233 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a406cec-8381-4a82-8d46-3ababcda72b4-config\") pod \"6a406cec-8381-4a82-8d46-3ababcda72b4\" (UID: \"6a406cec-8381-4a82-8d46-3ababcda72b4\") " Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.752304 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d2wq\" (UniqueName: \"kubernetes.io/projected/6a406cec-8381-4a82-8d46-3ababcda72b4-kube-api-access-8d2wq\") pod \"6a406cec-8381-4a82-8d46-3ababcda72b4\" (UID: \"6a406cec-8381-4a82-8d46-3ababcda72b4\") " Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.753315 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a406cec-8381-4a82-8d46-3ababcda72b4-client-ca" (OuterVolumeSpecName: "client-ca") pod "6a406cec-8381-4a82-8d46-3ababcda72b4" (UID: "6a406cec-8381-4a82-8d46-3ababcda72b4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.753539 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a406cec-8381-4a82-8d46-3ababcda72b4-config" (OuterVolumeSpecName: "config") pod "6a406cec-8381-4a82-8d46-3ababcda72b4" (UID: "6a406cec-8381-4a82-8d46-3ababcda72b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.771188 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a406cec-8381-4a82-8d46-3ababcda72b4-kube-api-access-8d2wq" (OuterVolumeSpecName: "kube-api-access-8d2wq") pod "6a406cec-8381-4a82-8d46-3ababcda72b4" (UID: "6a406cec-8381-4a82-8d46-3ababcda72b4"). InnerVolumeSpecName "kube-api-access-8d2wq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.772222 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a406cec-8381-4a82-8d46-3ababcda72b4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6a406cec-8381-4a82-8d46-3ababcda72b4" (UID: "6a406cec-8381-4a82-8d46-3ababcda72b4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.853533 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-client-ca\") pod \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.853841 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-config\") pod \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.854000 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr2fg\" (UniqueName: \"kubernetes.io/projected/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-kube-api-access-wr2fg\") pod \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.854082 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-serving-cert\") pod \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.854156 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-client-ca" (OuterVolumeSpecName: "client-ca") pod "a8fc6bf6-a48e-4f93-8a6c-57614d1171b7" (UID: "a8fc6bf6-a48e-4f93-8a6c-57614d1171b7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.854263 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-proxy-ca-bundles\") pod \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\" (UID: \"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7\") " Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.854617 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d2wq\" (UniqueName: \"kubernetes.io/projected/6a406cec-8381-4a82-8d46-3ababcda72b4-kube-api-access-8d2wq\") on node \"crc\" DevicePath \"\"" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.854683 4818 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.854745 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a406cec-8381-4a82-8d46-3ababcda72b4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.854979 4818 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a406cec-8381-4a82-8d46-3ababcda72b4-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.855042 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a406cec-8381-4a82-8d46-3ababcda72b4-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.855058 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a8fc6bf6-a48e-4f93-8a6c-57614d1171b7" (UID: "a8fc6bf6-a48e-4f93-8a6c-57614d1171b7"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.855556 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-config" (OuterVolumeSpecName: "config") pod "a8fc6bf6-a48e-4f93-8a6c-57614d1171b7" (UID: "a8fc6bf6-a48e-4f93-8a6c-57614d1171b7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.857304 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a8fc6bf6-a48e-4f93-8a6c-57614d1171b7" (UID: "a8fc6bf6-a48e-4f93-8a6c-57614d1171b7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.857940 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-kube-api-access-wr2fg" (OuterVolumeSpecName: "kube-api-access-wr2fg") pod "a8fc6bf6-a48e-4f93-8a6c-57614d1171b7" (UID: "a8fc6bf6-a48e-4f93-8a6c-57614d1171b7"). InnerVolumeSpecName "kube-api-access-wr2fg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.956512 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.956548 4818 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.956561 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:32:54 crc kubenswrapper[4818]: I1203 06:32:54.956572 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr2fg\" (UniqueName: \"kubernetes.io/projected/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7-kube-api-access-wr2fg\") on node \"crc\" DevicePath \"\"" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.436477 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" event={"ID":"6a406cec-8381-4a82-8d46-3ababcda72b4","Type":"ContainerDied","Data":"e8006f24ea9ac229abfd1f35c197ad198a1e9e8c896924f87daba48046e1d43d"} Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.436593 4818 scope.go:117] "RemoveContainer" containerID="8941867e59b8f0abdf51e22f253e435eea713ca60abc759608fac0bbc364f02a" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.436956 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.442106 4818 generic.go:334] "Generic (PLEG): container finished" podID="a8fc6bf6-a48e-4f93-8a6c-57614d1171b7" containerID="37f806687d01927c0c9fe9c57dd2e12b9ce2f0a32b8287d06ea0ca1993b5d98b" exitCode=0 Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.442158 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" event={"ID":"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7","Type":"ContainerDied","Data":"37f806687d01927c0c9fe9c57dd2e12b9ce2f0a32b8287d06ea0ca1993b5d98b"} Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.442212 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" event={"ID":"a8fc6bf6-a48e-4f93-8a6c-57614d1171b7","Type":"ContainerDied","Data":"86a0a80e10db78e3141bdd3d5232448b580b394eedcf1c04b18d7d80ebc41fc5"} Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.442244 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-cwhmk" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.466462 4818 scope.go:117] "RemoveContainer" containerID="37f806687d01927c0c9fe9c57dd2e12b9ce2f0a32b8287d06ea0ca1993b5d98b" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.493199 4818 scope.go:117] "RemoveContainer" containerID="37f806687d01927c0c9fe9c57dd2e12b9ce2f0a32b8287d06ea0ca1993b5d98b" Dec 03 06:32:55 crc kubenswrapper[4818]: E1203 06:32:55.493906 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37f806687d01927c0c9fe9c57dd2e12b9ce2f0a32b8287d06ea0ca1993b5d98b\": container with ID starting with 37f806687d01927c0c9fe9c57dd2e12b9ce2f0a32b8287d06ea0ca1993b5d98b not found: ID does not exist" containerID="37f806687d01927c0c9fe9c57dd2e12b9ce2f0a32b8287d06ea0ca1993b5d98b" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.493980 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37f806687d01927c0c9fe9c57dd2e12b9ce2f0a32b8287d06ea0ca1993b5d98b"} err="failed to get container status \"37f806687d01927c0c9fe9c57dd2e12b9ce2f0a32b8287d06ea0ca1993b5d98b\": rpc error: code = NotFound desc = could not find container \"37f806687d01927c0c9fe9c57dd2e12b9ce2f0a32b8287d06ea0ca1993b5d98b\": container with ID starting with 37f806687d01927c0c9fe9c57dd2e12b9ce2f0a32b8287d06ea0ca1993b5d98b not found: ID does not exist" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.499763 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58"] Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.507670 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5jw58"] Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.520044 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cwhmk"] Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.526687 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-cwhmk"] Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.971828 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-df87b5455-56gwp"] Dec 03 06:32:55 crc kubenswrapper[4818]: E1203 06:32:55.972303 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" containerName="installer" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.972367 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" containerName="installer" Dec 03 06:32:55 crc kubenswrapper[4818]: E1203 06:32:55.972382 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.972387 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 06:32:55 crc kubenswrapper[4818]: E1203 06:32:55.972400 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a406cec-8381-4a82-8d46-3ababcda72b4" containerName="route-controller-manager" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.972406 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a406cec-8381-4a82-8d46-3ababcda72b4" containerName="route-controller-manager" Dec 03 06:32:55 crc kubenswrapper[4818]: E1203 06:32:55.972419 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8fc6bf6-a48e-4f93-8a6c-57614d1171b7" containerName="controller-manager" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.972427 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8fc6bf6-a48e-4f93-8a6c-57614d1171b7" containerName="controller-manager" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.972527 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="24a1ab03-bbf6-4d53-9adb-ddc42b90fdde" containerName="installer" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.972539 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.972553 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8fc6bf6-a48e-4f93-8a6c-57614d1171b7" containerName="controller-manager" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.972561 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a406cec-8381-4a82-8d46-3ababcda72b4" containerName="route-controller-manager" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.973015 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.975501 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.975512 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.976038 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.977553 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.980059 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.982909 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.989722 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-df87b5455-56gwp"] Dec 03 06:32:55 crc kubenswrapper[4818]: I1203 06:32:55.990855 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.127185 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84"] Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.128250 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.133171 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.133402 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.133646 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.133869 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.134019 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.135037 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.142833 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84"] Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.171342 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrwg9\" (UniqueName: \"kubernetes.io/projected/d4a8bc0c-8ec0-410e-a4fb-710f31b9345e-kube-api-access-vrwg9\") pod \"controller-manager-df87b5455-56gwp\" (UID: \"d4a8bc0c-8ec0-410e-a4fb-710f31b9345e\") " pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.171418 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4a8bc0c-8ec0-410e-a4fb-710f31b9345e-proxy-ca-bundles\") pod \"controller-manager-df87b5455-56gwp\" (UID: \"d4a8bc0c-8ec0-410e-a4fb-710f31b9345e\") " pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.171621 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4a8bc0c-8ec0-410e-a4fb-710f31b9345e-serving-cert\") pod \"controller-manager-df87b5455-56gwp\" (UID: \"d4a8bc0c-8ec0-410e-a4fb-710f31b9345e\") " pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.171658 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4a8bc0c-8ec0-410e-a4fb-710f31b9345e-config\") pod \"controller-manager-df87b5455-56gwp\" (UID: \"d4a8bc0c-8ec0-410e-a4fb-710f31b9345e\") " pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.171692 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4a8bc0c-8ec0-410e-a4fb-710f31b9345e-client-ca\") pod \"controller-manager-df87b5455-56gwp\" (UID: \"d4a8bc0c-8ec0-410e-a4fb-710f31b9345e\") " pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.273080 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrwg9\" (UniqueName: \"kubernetes.io/projected/d4a8bc0c-8ec0-410e-a4fb-710f31b9345e-kube-api-access-vrwg9\") pod \"controller-manager-df87b5455-56gwp\" (UID: \"d4a8bc0c-8ec0-410e-a4fb-710f31b9345e\") " pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.273141 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2c6s\" (UniqueName: \"kubernetes.io/projected/7cfbb384-1524-46bc-b6f7-927326a7f8c3-kube-api-access-k2c6s\") pod \"route-controller-manager-6b5d4bcb95-54g84\" (UID: \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.273178 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7cfbb384-1524-46bc-b6f7-927326a7f8c3-client-ca\") pod \"route-controller-manager-6b5d4bcb95-54g84\" (UID: \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.273203 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4a8bc0c-8ec0-410e-a4fb-710f31b9345e-proxy-ca-bundles\") pod \"controller-manager-df87b5455-56gwp\" (UID: \"d4a8bc0c-8ec0-410e-a4fb-710f31b9345e\") " pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.273255 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4a8bc0c-8ec0-410e-a4fb-710f31b9345e-serving-cert\") pod \"controller-manager-df87b5455-56gwp\" (UID: \"d4a8bc0c-8ec0-410e-a4fb-710f31b9345e\") " pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.273292 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4a8bc0c-8ec0-410e-a4fb-710f31b9345e-config\") pod \"controller-manager-df87b5455-56gwp\" (UID: \"d4a8bc0c-8ec0-410e-a4fb-710f31b9345e\") " pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.273318 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cfbb384-1524-46bc-b6f7-927326a7f8c3-config\") pod \"route-controller-manager-6b5d4bcb95-54g84\" (UID: \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.273341 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cfbb384-1524-46bc-b6f7-927326a7f8c3-serving-cert\") pod \"route-controller-manager-6b5d4bcb95-54g84\" (UID: \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.273364 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4a8bc0c-8ec0-410e-a4fb-710f31b9345e-client-ca\") pod \"controller-manager-df87b5455-56gwp\" (UID: \"d4a8bc0c-8ec0-410e-a4fb-710f31b9345e\") " pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.274240 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4a8bc0c-8ec0-410e-a4fb-710f31b9345e-client-ca\") pod \"controller-manager-df87b5455-56gwp\" (UID: \"d4a8bc0c-8ec0-410e-a4fb-710f31b9345e\") " pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.274601 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4a8bc0c-8ec0-410e-a4fb-710f31b9345e-proxy-ca-bundles\") pod \"controller-manager-df87b5455-56gwp\" (UID: \"d4a8bc0c-8ec0-410e-a4fb-710f31b9345e\") " pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.274618 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4a8bc0c-8ec0-410e-a4fb-710f31b9345e-config\") pod \"controller-manager-df87b5455-56gwp\" (UID: \"d4a8bc0c-8ec0-410e-a4fb-710f31b9345e\") " pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.289763 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4a8bc0c-8ec0-410e-a4fb-710f31b9345e-serving-cert\") pod \"controller-manager-df87b5455-56gwp\" (UID: \"d4a8bc0c-8ec0-410e-a4fb-710f31b9345e\") " pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.293195 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrwg9\" (UniqueName: \"kubernetes.io/projected/d4a8bc0c-8ec0-410e-a4fb-710f31b9345e-kube-api-access-vrwg9\") pod \"controller-manager-df87b5455-56gwp\" (UID: \"d4a8bc0c-8ec0-410e-a4fb-710f31b9345e\") " pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.347939 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.374155 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2c6s\" (UniqueName: \"kubernetes.io/projected/7cfbb384-1524-46bc-b6f7-927326a7f8c3-kube-api-access-k2c6s\") pod \"route-controller-manager-6b5d4bcb95-54g84\" (UID: \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.374358 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7cfbb384-1524-46bc-b6f7-927326a7f8c3-client-ca\") pod \"route-controller-manager-6b5d4bcb95-54g84\" (UID: \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.374450 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cfbb384-1524-46bc-b6f7-927326a7f8c3-config\") pod \"route-controller-manager-6b5d4bcb95-54g84\" (UID: \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.374531 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cfbb384-1524-46bc-b6f7-927326a7f8c3-serving-cert\") pod \"route-controller-manager-6b5d4bcb95-54g84\" (UID: \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.375530 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7cfbb384-1524-46bc-b6f7-927326a7f8c3-client-ca\") pod \"route-controller-manager-6b5d4bcb95-54g84\" (UID: \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.375557 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cfbb384-1524-46bc-b6f7-927326a7f8c3-config\") pod \"route-controller-manager-6b5d4bcb95-54g84\" (UID: \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.378233 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cfbb384-1524-46bc-b6f7-927326a7f8c3-serving-cert\") pod \"route-controller-manager-6b5d4bcb95-54g84\" (UID: \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.393649 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2c6s\" (UniqueName: \"kubernetes.io/projected/7cfbb384-1524-46bc-b6f7-927326a7f8c3-kube-api-access-k2c6s\") pod \"route-controller-manager-6b5d4bcb95-54g84\" (UID: \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\") " pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.449876 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.522735 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-df87b5455-56gwp"] Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.635956 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84"] Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.749833 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a406cec-8381-4a82-8d46-3ababcda72b4" path="/var/lib/kubelet/pods/6a406cec-8381-4a82-8d46-3ababcda72b4/volumes" Dec 03 06:32:56 crc kubenswrapper[4818]: I1203 06:32:56.750766 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8fc6bf6-a48e-4f93-8a6c-57614d1171b7" path="/var/lib/kubelet/pods/a8fc6bf6-a48e-4f93-8a6c-57614d1171b7/volumes" Dec 03 06:32:57 crc kubenswrapper[4818]: I1203 06:32:57.478588 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" event={"ID":"d4a8bc0c-8ec0-410e-a4fb-710f31b9345e","Type":"ContainerStarted","Data":"a3858c2518a96ace6b8083d1b6badccd1f9cc282132279a721af6d46054f59c9"} Dec 03 06:32:57 crc kubenswrapper[4818]: I1203 06:32:57.479915 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" event={"ID":"d4a8bc0c-8ec0-410e-a4fb-710f31b9345e","Type":"ContainerStarted","Data":"f9c1b31fe2d91f1600b5434094357a795ea2a74d18fdbb89fda06ee8cf90772a"} Dec 03 06:32:57 crc kubenswrapper[4818]: I1203 06:32:57.480034 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:57 crc kubenswrapper[4818]: I1203 06:32:57.481110 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" event={"ID":"7cfbb384-1524-46bc-b6f7-927326a7f8c3","Type":"ContainerStarted","Data":"64858c15c350f80445bdd0bf1d4c79218a03f3fcd11b3bef87eaa80bb5d4d056"} Dec 03 06:32:57 crc kubenswrapper[4818]: I1203 06:32:57.481166 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" event={"ID":"7cfbb384-1524-46bc-b6f7-927326a7f8c3","Type":"ContainerStarted","Data":"6b33f9cc3b6ec4c1ababea14a870f12a95bfea27718f9327fb46eaff0fba8cd5"} Dec 03 06:32:57 crc kubenswrapper[4818]: I1203 06:32:57.481428 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:32:57 crc kubenswrapper[4818]: I1203 06:32:57.486177 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" Dec 03 06:32:57 crc kubenswrapper[4818]: I1203 06:32:57.492594 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:32:57 crc kubenswrapper[4818]: I1203 06:32:57.521542 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-df87b5455-56gwp" podStartSLOduration=2.521524102 podStartE2EDuration="2.521524102s" podCreationTimestamp="2025-12-03 06:32:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:32:57.50496447 +0000 UTC m=+335.196573222" watchObservedRunningTime="2025-12-03 06:32:57.521524102 +0000 UTC m=+335.213132854" Dec 03 06:32:57 crc kubenswrapper[4818]: I1203 06:32:57.522941 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" podStartSLOduration=3.522932406 podStartE2EDuration="3.522932406s" podCreationTimestamp="2025-12-03 06:32:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:32:57.520781964 +0000 UTC m=+335.212390716" watchObservedRunningTime="2025-12-03 06:32:57.522932406 +0000 UTC m=+335.214541158" Dec 03 06:33:13 crc kubenswrapper[4818]: I1203 06:33:13.302589 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:33:13 crc kubenswrapper[4818]: I1203 06:33:13.303659 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.057317 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84"] Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.057886 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" podUID="7cfbb384-1524-46bc-b6f7-927326a7f8c3" containerName="route-controller-manager" containerID="cri-o://64858c15c350f80445bdd0bf1d4c79218a03f3fcd11b3bef87eaa80bb5d4d056" gracePeriod=30 Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.493238 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.522070 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7cfbb384-1524-46bc-b6f7-927326a7f8c3-client-ca\") pod \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\" (UID: \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\") " Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.522171 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cfbb384-1524-46bc-b6f7-927326a7f8c3-config\") pod \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\" (UID: \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\") " Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.522243 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cfbb384-1524-46bc-b6f7-927326a7f8c3-serving-cert\") pod \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\" (UID: \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\") " Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.522360 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2c6s\" (UniqueName: \"kubernetes.io/projected/7cfbb384-1524-46bc-b6f7-927326a7f8c3-kube-api-access-k2c6s\") pod \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\" (UID: \"7cfbb384-1524-46bc-b6f7-927326a7f8c3\") " Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.523331 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cfbb384-1524-46bc-b6f7-927326a7f8c3-config" (OuterVolumeSpecName: "config") pod "7cfbb384-1524-46bc-b6f7-927326a7f8c3" (UID: "7cfbb384-1524-46bc-b6f7-927326a7f8c3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.523294 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cfbb384-1524-46bc-b6f7-927326a7f8c3-client-ca" (OuterVolumeSpecName: "client-ca") pod "7cfbb384-1524-46bc-b6f7-927326a7f8c3" (UID: "7cfbb384-1524-46bc-b6f7-927326a7f8c3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.523534 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7cfbb384-1524-46bc-b6f7-927326a7f8c3-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.523573 4818 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7cfbb384-1524-46bc-b6f7-927326a7f8c3-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.529499 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cfbb384-1524-46bc-b6f7-927326a7f8c3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7cfbb384-1524-46bc-b6f7-927326a7f8c3" (UID: "7cfbb384-1524-46bc-b6f7-927326a7f8c3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.534080 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cfbb384-1524-46bc-b6f7-927326a7f8c3-kube-api-access-k2c6s" (OuterVolumeSpecName: "kube-api-access-k2c6s") pod "7cfbb384-1524-46bc-b6f7-927326a7f8c3" (UID: "7cfbb384-1524-46bc-b6f7-927326a7f8c3"). InnerVolumeSpecName "kube-api-access-k2c6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.580382 4818 generic.go:334] "Generic (PLEG): container finished" podID="7cfbb384-1524-46bc-b6f7-927326a7f8c3" containerID="64858c15c350f80445bdd0bf1d4c79218a03f3fcd11b3bef87eaa80bb5d4d056" exitCode=0 Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.580459 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" event={"ID":"7cfbb384-1524-46bc-b6f7-927326a7f8c3","Type":"ContainerDied","Data":"64858c15c350f80445bdd0bf1d4c79218a03f3fcd11b3bef87eaa80bb5d4d056"} Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.580530 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.580566 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84" event={"ID":"7cfbb384-1524-46bc-b6f7-927326a7f8c3","Type":"ContainerDied","Data":"6b33f9cc3b6ec4c1ababea14a870f12a95bfea27718f9327fb46eaff0fba8cd5"} Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.580590 4818 scope.go:117] "RemoveContainer" containerID="64858c15c350f80445bdd0bf1d4c79218a03f3fcd11b3bef87eaa80bb5d4d056" Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.603591 4818 scope.go:117] "RemoveContainer" containerID="64858c15c350f80445bdd0bf1d4c79218a03f3fcd11b3bef87eaa80bb5d4d056" Dec 03 06:33:14 crc kubenswrapper[4818]: E1203 06:33:14.604114 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64858c15c350f80445bdd0bf1d4c79218a03f3fcd11b3bef87eaa80bb5d4d056\": container with ID starting with 64858c15c350f80445bdd0bf1d4c79218a03f3fcd11b3bef87eaa80bb5d4d056 not found: ID does not exist" containerID="64858c15c350f80445bdd0bf1d4c79218a03f3fcd11b3bef87eaa80bb5d4d056" Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.604166 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64858c15c350f80445bdd0bf1d4c79218a03f3fcd11b3bef87eaa80bb5d4d056"} err="failed to get container status \"64858c15c350f80445bdd0bf1d4c79218a03f3fcd11b3bef87eaa80bb5d4d056\": rpc error: code = NotFound desc = could not find container \"64858c15c350f80445bdd0bf1d4c79218a03f3fcd11b3bef87eaa80bb5d4d056\": container with ID starting with 64858c15c350f80445bdd0bf1d4c79218a03f3fcd11b3bef87eaa80bb5d4d056 not found: ID does not exist" Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.620784 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84"] Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.623841 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5d4bcb95-54g84"] Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.624845 4818 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7cfbb384-1524-46bc-b6f7-927326a7f8c3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.624878 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2c6s\" (UniqueName: \"kubernetes.io/projected/7cfbb384-1524-46bc-b6f7-927326a7f8c3-kube-api-access-k2c6s\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:14 crc kubenswrapper[4818]: I1203 06:33:14.747092 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cfbb384-1524-46bc-b6f7-927326a7f8c3" path="/var/lib/kubelet/pods/7cfbb384-1524-46bc-b6f7-927326a7f8c3/volumes" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.153921 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k"] Dec 03 06:33:15 crc kubenswrapper[4818]: E1203 06:33:15.154271 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cfbb384-1524-46bc-b6f7-927326a7f8c3" containerName="route-controller-manager" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.154296 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cfbb384-1524-46bc-b6f7-927326a7f8c3" containerName="route-controller-manager" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.154480 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cfbb384-1524-46bc-b6f7-927326a7f8c3" containerName="route-controller-manager" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.155067 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.158949 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.159231 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.159628 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.159851 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.160016 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.161143 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.169381 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k"] Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.233636 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb59r\" (UniqueName: \"kubernetes.io/projected/5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a-kube-api-access-sb59r\") pod \"route-controller-manager-55dbd88cdb-z2g5k\" (UID: \"5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a\") " pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.233771 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a-client-ca\") pod \"route-controller-manager-55dbd88cdb-z2g5k\" (UID: \"5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a\") " pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.233866 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a-serving-cert\") pod \"route-controller-manager-55dbd88cdb-z2g5k\" (UID: \"5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a\") " pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.234263 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a-config\") pod \"route-controller-manager-55dbd88cdb-z2g5k\" (UID: \"5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a\") " pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.335539 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a-serving-cert\") pod \"route-controller-manager-55dbd88cdb-z2g5k\" (UID: \"5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a\") " pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.335655 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a-config\") pod \"route-controller-manager-55dbd88cdb-z2g5k\" (UID: \"5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a\") " pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.335690 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb59r\" (UniqueName: \"kubernetes.io/projected/5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a-kube-api-access-sb59r\") pod \"route-controller-manager-55dbd88cdb-z2g5k\" (UID: \"5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a\") " pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.335735 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a-client-ca\") pod \"route-controller-manager-55dbd88cdb-z2g5k\" (UID: \"5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a\") " pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.336983 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a-client-ca\") pod \"route-controller-manager-55dbd88cdb-z2g5k\" (UID: \"5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a\") " pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.337230 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a-config\") pod \"route-controller-manager-55dbd88cdb-z2g5k\" (UID: \"5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a\") " pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.349619 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a-serving-cert\") pod \"route-controller-manager-55dbd88cdb-z2g5k\" (UID: \"5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a\") " pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.355178 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb59r\" (UniqueName: \"kubernetes.io/projected/5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a-kube-api-access-sb59r\") pod \"route-controller-manager-55dbd88cdb-z2g5k\" (UID: \"5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a\") " pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.476684 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" Dec 03 06:33:15 crc kubenswrapper[4818]: I1203 06:33:15.889250 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k"] Dec 03 06:33:15 crc kubenswrapper[4818]: W1203 06:33:15.898420 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c1ae9b4_e30a_473b_aa20_b5eacbe7b43a.slice/crio-dbf7f98db5ab00a9bcdf28387ff3500970ec723a81913ab8adcb19c94076f3f0 WatchSource:0}: Error finding container dbf7f98db5ab00a9bcdf28387ff3500970ec723a81913ab8adcb19c94076f3f0: Status 404 returned error can't find the container with id dbf7f98db5ab00a9bcdf28387ff3500970ec723a81913ab8adcb19c94076f3f0 Dec 03 06:33:16 crc kubenswrapper[4818]: I1203 06:33:16.596995 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" event={"ID":"5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a","Type":"ContainerStarted","Data":"9795e3be31f3e9704de89fd2687b64033af120e3bf0a9454b86ebed8afdc53b7"} Dec 03 06:33:16 crc kubenswrapper[4818]: I1203 06:33:16.597355 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" event={"ID":"5c1ae9b4-e30a-473b-aa20-b5eacbe7b43a","Type":"ContainerStarted","Data":"dbf7f98db5ab00a9bcdf28387ff3500970ec723a81913ab8adcb19c94076f3f0"} Dec 03 06:33:16 crc kubenswrapper[4818]: I1203 06:33:16.598839 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" Dec 03 06:33:16 crc kubenswrapper[4818]: I1203 06:33:16.623770 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" podStartSLOduration=2.623740196 podStartE2EDuration="2.623740196s" podCreationTimestamp="2025-12-03 06:33:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:33:16.620797624 +0000 UTC m=+354.312406416" watchObservedRunningTime="2025-12-03 06:33:16.623740196 +0000 UTC m=+354.315348948" Dec 03 06:33:16 crc kubenswrapper[4818]: I1203 06:33:16.644936 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-55dbd88cdb-z2g5k" Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.596652 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jrfsv"] Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.598021 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jrfsv" podUID="0c4ffde8-459a-4634-b567-5f900d40ceec" containerName="registry-server" containerID="cri-o://1ea4fe4551a71e496e406f333497c69daf15b0f43c41a115de3c6500e2520da6" gracePeriod=30 Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.627124 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k7wwb"] Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.627427 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k7wwb" podUID="0301f892-ca1d-4f8b-a429-6d0ad96a50dc" containerName="registry-server" containerID="cri-o://04caaf0e08ce8e688a22ab83ddfb7237545bbbb93f5e378ad7d0955d1d9446fd" gracePeriod=30 Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.634184 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4dh8d"] Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.634416 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" podUID="53a413a7-34f8-4fb3-893b-c4fa9711017a" containerName="marketplace-operator" containerID="cri-o://3395f8e86e03b477f14f091b8e9c92d7de78b57f50e6601bb62d2c7e1dbad671" gracePeriod=30 Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.654367 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hl66m"] Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.655301 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hl66m" Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.668391 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-965hr"] Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.669023 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-965hr" podUID="91223086-535a-47c4-a843-3bb100a7da50" containerName="registry-server" containerID="cri-o://3e37f9cc5816052f7c6991762135723f0f86e6f484e32bd8e2870d46593d55bc" gracePeriod=30 Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.678437 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x7vlw"] Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.678701 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x7vlw" podUID="75955bfe-ae6e-481d-84b5-3de486a86827" containerName="registry-server" containerID="cri-o://8e606076dd130f5dbe3724a712bad6f18f49e260d7aadf33a5b803edde0431f8" gracePeriod=30 Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.684436 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hl66m"] Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.750030 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96b4v\" (UniqueName: \"kubernetes.io/projected/a89a1dfa-1797-48f8-bd2d-f968c523e8df-kube-api-access-96b4v\") pod \"marketplace-operator-79b997595-hl66m\" (UID: \"a89a1dfa-1797-48f8-bd2d-f968c523e8df\") " pod="openshift-marketplace/marketplace-operator-79b997595-hl66m" Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.750139 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a89a1dfa-1797-48f8-bd2d-f968c523e8df-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hl66m\" (UID: \"a89a1dfa-1797-48f8-bd2d-f968c523e8df\") " pod="openshift-marketplace/marketplace-operator-79b997595-hl66m" Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.750176 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a89a1dfa-1797-48f8-bd2d-f968c523e8df-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hl66m\" (UID: \"a89a1dfa-1797-48f8-bd2d-f968c523e8df\") " pod="openshift-marketplace/marketplace-operator-79b997595-hl66m" Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.851376 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a89a1dfa-1797-48f8-bd2d-f968c523e8df-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hl66m\" (UID: \"a89a1dfa-1797-48f8-bd2d-f968c523e8df\") " pod="openshift-marketplace/marketplace-operator-79b997595-hl66m" Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.851440 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a89a1dfa-1797-48f8-bd2d-f968c523e8df-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hl66m\" (UID: \"a89a1dfa-1797-48f8-bd2d-f968c523e8df\") " pod="openshift-marketplace/marketplace-operator-79b997595-hl66m" Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.851596 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96b4v\" (UniqueName: \"kubernetes.io/projected/a89a1dfa-1797-48f8-bd2d-f968c523e8df-kube-api-access-96b4v\") pod \"marketplace-operator-79b997595-hl66m\" (UID: \"a89a1dfa-1797-48f8-bd2d-f968c523e8df\") " pod="openshift-marketplace/marketplace-operator-79b997595-hl66m" Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.853280 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a89a1dfa-1797-48f8-bd2d-f968c523e8df-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hl66m\" (UID: \"a89a1dfa-1797-48f8-bd2d-f968c523e8df\") " pod="openshift-marketplace/marketplace-operator-79b997595-hl66m" Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.858016 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a89a1dfa-1797-48f8-bd2d-f968c523e8df-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hl66m\" (UID: \"a89a1dfa-1797-48f8-bd2d-f968c523e8df\") " pod="openshift-marketplace/marketplace-operator-79b997595-hl66m" Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.870420 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96b4v\" (UniqueName: \"kubernetes.io/projected/a89a1dfa-1797-48f8-bd2d-f968c523e8df-kube-api-access-96b4v\") pod \"marketplace-operator-79b997595-hl66m\" (UID: \"a89a1dfa-1797-48f8-bd2d-f968c523e8df\") " pod="openshift-marketplace/marketplace-operator-79b997595-hl66m" Dec 03 06:33:30 crc kubenswrapper[4818]: I1203 06:33:30.977985 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hl66m" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.095203 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jrfsv" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.182661 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-965hr" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.186220 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.191970 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k7wwb" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.194858 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7vlw" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.261862 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-879pq\" (UniqueName: \"kubernetes.io/projected/0c4ffde8-459a-4634-b567-5f900d40ceec-kube-api-access-879pq\") pod \"0c4ffde8-459a-4634-b567-5f900d40ceec\" (UID: \"0c4ffde8-459a-4634-b567-5f900d40ceec\") " Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.261915 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c4ffde8-459a-4634-b567-5f900d40ceec-catalog-content\") pod \"0c4ffde8-459a-4634-b567-5f900d40ceec\" (UID: \"0c4ffde8-459a-4634-b567-5f900d40ceec\") " Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.261963 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c4ffde8-459a-4634-b567-5f900d40ceec-utilities\") pod \"0c4ffde8-459a-4634-b567-5f900d40ceec\" (UID: \"0c4ffde8-459a-4634-b567-5f900d40ceec\") " Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.263269 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c4ffde8-459a-4634-b567-5f900d40ceec-utilities" (OuterVolumeSpecName: "utilities") pod "0c4ffde8-459a-4634-b567-5f900d40ceec" (UID: "0c4ffde8-459a-4634-b567-5f900d40ceec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.267558 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c4ffde8-459a-4634-b567-5f900d40ceec-kube-api-access-879pq" (OuterVolumeSpecName: "kube-api-access-879pq") pod "0c4ffde8-459a-4634-b567-5f900d40ceec" (UID: "0c4ffde8-459a-4634-b567-5f900d40ceec"). InnerVolumeSpecName "kube-api-access-879pq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.335637 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c4ffde8-459a-4634-b567-5f900d40ceec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c4ffde8-459a-4634-b567-5f900d40ceec" (UID: "0c4ffde8-459a-4634-b567-5f900d40ceec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.363440 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tscsg\" (UniqueName: \"kubernetes.io/projected/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-kube-api-access-tscsg\") pod \"0301f892-ca1d-4f8b-a429-6d0ad96a50dc\" (UID: \"0301f892-ca1d-4f8b-a429-6d0ad96a50dc\") " Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.363533 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75955bfe-ae6e-481d-84b5-3de486a86827-utilities\") pod \"75955bfe-ae6e-481d-84b5-3de486a86827\" (UID: \"75955bfe-ae6e-481d-84b5-3de486a86827\") " Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.363559 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91223086-535a-47c4-a843-3bb100a7da50-catalog-content\") pod \"91223086-535a-47c4-a843-3bb100a7da50\" (UID: \"91223086-535a-47c4-a843-3bb100a7da50\") " Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.363591 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw8b8\" (UniqueName: \"kubernetes.io/projected/91223086-535a-47c4-a843-3bb100a7da50-kube-api-access-tw8b8\") pod \"91223086-535a-47c4-a843-3bb100a7da50\" (UID: \"91223086-535a-47c4-a843-3bb100a7da50\") " Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.363637 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91223086-535a-47c4-a843-3bb100a7da50-utilities\") pod \"91223086-535a-47c4-a843-3bb100a7da50\" (UID: \"91223086-535a-47c4-a843-3bb100a7da50\") " Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.363681 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-utilities\") pod \"0301f892-ca1d-4f8b-a429-6d0ad96a50dc\" (UID: \"0301f892-ca1d-4f8b-a429-6d0ad96a50dc\") " Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.363709 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53a413a7-34f8-4fb3-893b-c4fa9711017a-marketplace-trusted-ca\") pod \"53a413a7-34f8-4fb3-893b-c4fa9711017a\" (UID: \"53a413a7-34f8-4fb3-893b-c4fa9711017a\") " Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.364764 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75955bfe-ae6e-481d-84b5-3de486a86827-catalog-content\") pod \"75955bfe-ae6e-481d-84b5-3de486a86827\" (UID: \"75955bfe-ae6e-481d-84b5-3de486a86827\") " Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.364850 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-catalog-content\") pod \"0301f892-ca1d-4f8b-a429-6d0ad96a50dc\" (UID: \"0301f892-ca1d-4f8b-a429-6d0ad96a50dc\") " Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.364897 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p2fj\" (UniqueName: \"kubernetes.io/projected/75955bfe-ae6e-481d-84b5-3de486a86827-kube-api-access-6p2fj\") pod \"75955bfe-ae6e-481d-84b5-3de486a86827\" (UID: \"75955bfe-ae6e-481d-84b5-3de486a86827\") " Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.364943 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlm4m\" (UniqueName: \"kubernetes.io/projected/53a413a7-34f8-4fb3-893b-c4fa9711017a-kube-api-access-vlm4m\") pod \"53a413a7-34f8-4fb3-893b-c4fa9711017a\" (UID: \"53a413a7-34f8-4fb3-893b-c4fa9711017a\") " Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.365021 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/53a413a7-34f8-4fb3-893b-c4fa9711017a-marketplace-operator-metrics\") pod \"53a413a7-34f8-4fb3-893b-c4fa9711017a\" (UID: \"53a413a7-34f8-4fb3-893b-c4fa9711017a\") " Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.364936 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-utilities" (OuterVolumeSpecName: "utilities") pod "0301f892-ca1d-4f8b-a429-6d0ad96a50dc" (UID: "0301f892-ca1d-4f8b-a429-6d0ad96a50dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.364931 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91223086-535a-47c4-a843-3bb100a7da50-utilities" (OuterVolumeSpecName: "utilities") pod "91223086-535a-47c4-a843-3bb100a7da50" (UID: "91223086-535a-47c4-a843-3bb100a7da50"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.365317 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75955bfe-ae6e-481d-84b5-3de486a86827-utilities" (OuterVolumeSpecName: "utilities") pod "75955bfe-ae6e-481d-84b5-3de486a86827" (UID: "75955bfe-ae6e-481d-84b5-3de486a86827"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.366175 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-879pq\" (UniqueName: \"kubernetes.io/projected/0c4ffde8-459a-4634-b567-5f900d40ceec-kube-api-access-879pq\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.366245 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c4ffde8-459a-4634-b567-5f900d40ceec-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.366260 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75955bfe-ae6e-481d-84b5-3de486a86827-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.366277 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c4ffde8-459a-4634-b567-5f900d40ceec-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.366291 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91223086-535a-47c4-a843-3bb100a7da50-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.366304 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.366859 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91223086-535a-47c4-a843-3bb100a7da50-kube-api-access-tw8b8" (OuterVolumeSpecName: "kube-api-access-tw8b8") pod "91223086-535a-47c4-a843-3bb100a7da50" (UID: "91223086-535a-47c4-a843-3bb100a7da50"). InnerVolumeSpecName "kube-api-access-tw8b8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.367149 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-kube-api-access-tscsg" (OuterVolumeSpecName: "kube-api-access-tscsg") pod "0301f892-ca1d-4f8b-a429-6d0ad96a50dc" (UID: "0301f892-ca1d-4f8b-a429-6d0ad96a50dc"). InnerVolumeSpecName "kube-api-access-tscsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.368358 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75955bfe-ae6e-481d-84b5-3de486a86827-kube-api-access-6p2fj" (OuterVolumeSpecName: "kube-api-access-6p2fj") pod "75955bfe-ae6e-481d-84b5-3de486a86827" (UID: "75955bfe-ae6e-481d-84b5-3de486a86827"). InnerVolumeSpecName "kube-api-access-6p2fj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.368565 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53a413a7-34f8-4fb3-893b-c4fa9711017a-kube-api-access-vlm4m" (OuterVolumeSpecName: "kube-api-access-vlm4m") pod "53a413a7-34f8-4fb3-893b-c4fa9711017a" (UID: "53a413a7-34f8-4fb3-893b-c4fa9711017a"). InnerVolumeSpecName "kube-api-access-vlm4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.369544 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53a413a7-34f8-4fb3-893b-c4fa9711017a-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "53a413a7-34f8-4fb3-893b-c4fa9711017a" (UID: "53a413a7-34f8-4fb3-893b-c4fa9711017a"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.370070 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53a413a7-34f8-4fb3-893b-c4fa9711017a-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "53a413a7-34f8-4fb3-893b-c4fa9711017a" (UID: "53a413a7-34f8-4fb3-893b-c4fa9711017a"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.383837 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91223086-535a-47c4-a843-3bb100a7da50-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91223086-535a-47c4-a843-3bb100a7da50" (UID: "91223086-535a-47c4-a843-3bb100a7da50"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.412803 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0301f892-ca1d-4f8b-a429-6d0ad96a50dc" (UID: "0301f892-ca1d-4f8b-a429-6d0ad96a50dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.467920 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tscsg\" (UniqueName: \"kubernetes.io/projected/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-kube-api-access-tscsg\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.467955 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91223086-535a-47c4-a843-3bb100a7da50-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.467964 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw8b8\" (UniqueName: \"kubernetes.io/projected/91223086-535a-47c4-a843-3bb100a7da50-kube-api-access-tw8b8\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.467972 4818 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53a413a7-34f8-4fb3-893b-c4fa9711017a-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.467981 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0301f892-ca1d-4f8b-a429-6d0ad96a50dc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.467989 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p2fj\" (UniqueName: \"kubernetes.io/projected/75955bfe-ae6e-481d-84b5-3de486a86827-kube-api-access-6p2fj\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.467998 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlm4m\" (UniqueName: \"kubernetes.io/projected/53a413a7-34f8-4fb3-893b-c4fa9711017a-kube-api-access-vlm4m\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.468006 4818 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/53a413a7-34f8-4fb3-893b-c4fa9711017a-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.482352 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75955bfe-ae6e-481d-84b5-3de486a86827-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "75955bfe-ae6e-481d-84b5-3de486a86827" (UID: "75955bfe-ae6e-481d-84b5-3de486a86827"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.542089 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hl66m"] Dec 03 06:33:31 crc kubenswrapper[4818]: W1203 06:33:31.546969 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda89a1dfa_1797_48f8_bd2d_f968c523e8df.slice/crio-00faf08344dd8c2bd1d474c6414ffda1a2fe8d3e4423501618360ec07302496c WatchSource:0}: Error finding container 00faf08344dd8c2bd1d474c6414ffda1a2fe8d3e4423501618360ec07302496c: Status 404 returned error can't find the container with id 00faf08344dd8c2bd1d474c6414ffda1a2fe8d3e4423501618360ec07302496c Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.569097 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75955bfe-ae6e-481d-84b5-3de486a86827-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.698233 4818 generic.go:334] "Generic (PLEG): container finished" podID="75955bfe-ae6e-481d-84b5-3de486a86827" containerID="8e606076dd130f5dbe3724a712bad6f18f49e260d7aadf33a5b803edde0431f8" exitCode=0 Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.698330 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7vlw" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.698344 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7vlw" event={"ID":"75955bfe-ae6e-481d-84b5-3de486a86827","Type":"ContainerDied","Data":"8e606076dd130f5dbe3724a712bad6f18f49e260d7aadf33a5b803edde0431f8"} Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.698885 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7vlw" event={"ID":"75955bfe-ae6e-481d-84b5-3de486a86827","Type":"ContainerDied","Data":"fb107d86eb8bccc5f47d4ffdc35eaf554477a2a0592d8ae5d3629ab98e18e1f4"} Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.698920 4818 scope.go:117] "RemoveContainer" containerID="8e606076dd130f5dbe3724a712bad6f18f49e260d7aadf33a5b803edde0431f8" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.705662 4818 generic.go:334] "Generic (PLEG): container finished" podID="91223086-535a-47c4-a843-3bb100a7da50" containerID="3e37f9cc5816052f7c6991762135723f0f86e6f484e32bd8e2870d46593d55bc" exitCode=0 Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.705740 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-965hr" event={"ID":"91223086-535a-47c4-a843-3bb100a7da50","Type":"ContainerDied","Data":"3e37f9cc5816052f7c6991762135723f0f86e6f484e32bd8e2870d46593d55bc"} Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.705770 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-965hr" event={"ID":"91223086-535a-47c4-a843-3bb100a7da50","Type":"ContainerDied","Data":"3c57e09323739e67935be77ba6d4e194f10e027bfad4f009931b9eca389a8aa0"} Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.705780 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-965hr" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.713454 4818 scope.go:117] "RemoveContainer" containerID="893aac79870bf08760a3dadb13a02f9f0d3f8d8e4f7bdb9fc0176bd870851cda" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.715034 4818 generic.go:334] "Generic (PLEG): container finished" podID="0301f892-ca1d-4f8b-a429-6d0ad96a50dc" containerID="04caaf0e08ce8e688a22ab83ddfb7237545bbbb93f5e378ad7d0955d1d9446fd" exitCode=0 Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.715067 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7wwb" event={"ID":"0301f892-ca1d-4f8b-a429-6d0ad96a50dc","Type":"ContainerDied","Data":"04caaf0e08ce8e688a22ab83ddfb7237545bbbb93f5e378ad7d0955d1d9446fd"} Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.715114 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7wwb" event={"ID":"0301f892-ca1d-4f8b-a429-6d0ad96a50dc","Type":"ContainerDied","Data":"4625e08fc9cd6569bd4b04774a6c74dd03d178c2764532a534a5c5df30f3162c"} Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.715175 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k7wwb" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.719630 4818 generic.go:334] "Generic (PLEG): container finished" podID="53a413a7-34f8-4fb3-893b-c4fa9711017a" containerID="3395f8e86e03b477f14f091b8e9c92d7de78b57f50e6601bb62d2c7e1dbad671" exitCode=0 Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.719690 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" event={"ID":"53a413a7-34f8-4fb3-893b-c4fa9711017a","Type":"ContainerDied","Data":"3395f8e86e03b477f14f091b8e9c92d7de78b57f50e6601bb62d2c7e1dbad671"} Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.719719 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" event={"ID":"53a413a7-34f8-4fb3-893b-c4fa9711017a","Type":"ContainerDied","Data":"03df0cf759afe73c38c7041c610ba37cf8b523d7b5ed1ac477c6aa9c1a7c38f7"} Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.719800 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4dh8d" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.723977 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hl66m" event={"ID":"a89a1dfa-1797-48f8-bd2d-f968c523e8df","Type":"ContainerStarted","Data":"9a6a5d113511f522145e2174862f6350a9bd54de3f763f9a0e74ead25c94780c"} Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.724018 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hl66m" event={"ID":"a89a1dfa-1797-48f8-bd2d-f968c523e8df","Type":"ContainerStarted","Data":"00faf08344dd8c2bd1d474c6414ffda1a2fe8d3e4423501618360ec07302496c"} Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.724689 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-hl66m" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.726546 4818 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-hl66m container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.62:8080/healthz\": dial tcp 10.217.0.62:8080: connect: connection refused" start-of-body= Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.726598 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-hl66m" podUID="a89a1dfa-1797-48f8-bd2d-f968c523e8df" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.62:8080/healthz\": dial tcp 10.217.0.62:8080: connect: connection refused" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.735510 4818 generic.go:334] "Generic (PLEG): container finished" podID="0c4ffde8-459a-4634-b567-5f900d40ceec" containerID="1ea4fe4551a71e496e406f333497c69daf15b0f43c41a115de3c6500e2520da6" exitCode=0 Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.735584 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrfsv" event={"ID":"0c4ffde8-459a-4634-b567-5f900d40ceec","Type":"ContainerDied","Data":"1ea4fe4551a71e496e406f333497c69daf15b0f43c41a115de3c6500e2520da6"} Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.735626 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrfsv" event={"ID":"0c4ffde8-459a-4634-b567-5f900d40ceec","Type":"ContainerDied","Data":"a36a9c146ab72adba1f3bd6ba5135920ec58ce50478c9a9c11738cffafdcc878"} Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.735734 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jrfsv" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.740685 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x7vlw"] Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.741349 4818 scope.go:117] "RemoveContainer" containerID="4736d6686b2ae0fdb52c2f22ea23a0f6305b23829fb928b14edd686c66950d05" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.747291 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x7vlw"] Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.759168 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-965hr"] Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.763138 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-965hr"] Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.765219 4818 scope.go:117] "RemoveContainer" containerID="8e606076dd130f5dbe3724a712bad6f18f49e260d7aadf33a5b803edde0431f8" Dec 03 06:33:31 crc kubenswrapper[4818]: E1203 06:33:31.766346 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e606076dd130f5dbe3724a712bad6f18f49e260d7aadf33a5b803edde0431f8\": container with ID starting with 8e606076dd130f5dbe3724a712bad6f18f49e260d7aadf33a5b803edde0431f8 not found: ID does not exist" containerID="8e606076dd130f5dbe3724a712bad6f18f49e260d7aadf33a5b803edde0431f8" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.766394 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e606076dd130f5dbe3724a712bad6f18f49e260d7aadf33a5b803edde0431f8"} err="failed to get container status \"8e606076dd130f5dbe3724a712bad6f18f49e260d7aadf33a5b803edde0431f8\": rpc error: code = NotFound desc = could not find container \"8e606076dd130f5dbe3724a712bad6f18f49e260d7aadf33a5b803edde0431f8\": container with ID starting with 8e606076dd130f5dbe3724a712bad6f18f49e260d7aadf33a5b803edde0431f8 not found: ID does not exist" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.766431 4818 scope.go:117] "RemoveContainer" containerID="893aac79870bf08760a3dadb13a02f9f0d3f8d8e4f7bdb9fc0176bd870851cda" Dec 03 06:33:31 crc kubenswrapper[4818]: E1203 06:33:31.766797 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"893aac79870bf08760a3dadb13a02f9f0d3f8d8e4f7bdb9fc0176bd870851cda\": container with ID starting with 893aac79870bf08760a3dadb13a02f9f0d3f8d8e4f7bdb9fc0176bd870851cda not found: ID does not exist" containerID="893aac79870bf08760a3dadb13a02f9f0d3f8d8e4f7bdb9fc0176bd870851cda" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.766922 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"893aac79870bf08760a3dadb13a02f9f0d3f8d8e4f7bdb9fc0176bd870851cda"} err="failed to get container status \"893aac79870bf08760a3dadb13a02f9f0d3f8d8e4f7bdb9fc0176bd870851cda\": rpc error: code = NotFound desc = could not find container \"893aac79870bf08760a3dadb13a02f9f0d3f8d8e4f7bdb9fc0176bd870851cda\": container with ID starting with 893aac79870bf08760a3dadb13a02f9f0d3f8d8e4f7bdb9fc0176bd870851cda not found: ID does not exist" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.766962 4818 scope.go:117] "RemoveContainer" containerID="4736d6686b2ae0fdb52c2f22ea23a0f6305b23829fb928b14edd686c66950d05" Dec 03 06:33:31 crc kubenswrapper[4818]: E1203 06:33:31.768239 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4736d6686b2ae0fdb52c2f22ea23a0f6305b23829fb928b14edd686c66950d05\": container with ID starting with 4736d6686b2ae0fdb52c2f22ea23a0f6305b23829fb928b14edd686c66950d05 not found: ID does not exist" containerID="4736d6686b2ae0fdb52c2f22ea23a0f6305b23829fb928b14edd686c66950d05" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.770897 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4736d6686b2ae0fdb52c2f22ea23a0f6305b23829fb928b14edd686c66950d05"} err="failed to get container status \"4736d6686b2ae0fdb52c2f22ea23a0f6305b23829fb928b14edd686c66950d05\": rpc error: code = NotFound desc = could not find container \"4736d6686b2ae0fdb52c2f22ea23a0f6305b23829fb928b14edd686c66950d05\": container with ID starting with 4736d6686b2ae0fdb52c2f22ea23a0f6305b23829fb928b14edd686c66950d05 not found: ID does not exist" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.770985 4818 scope.go:117] "RemoveContainer" containerID="3e37f9cc5816052f7c6991762135723f0f86e6f484e32bd8e2870d46593d55bc" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.779047 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-hl66m" podStartSLOduration=1.778971924 podStartE2EDuration="1.778971924s" podCreationTimestamp="2025-12-03 06:33:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:33:31.775850128 +0000 UTC m=+369.467458880" watchObservedRunningTime="2025-12-03 06:33:31.778971924 +0000 UTC m=+369.470580676" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.808879 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k7wwb"] Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.816267 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k7wwb"] Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.823801 4818 scope.go:117] "RemoveContainer" containerID="8f0fe5cc139e0b9e63edec1a51c00eac2621dccc305d594aa83e600ea5c4212c" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.830590 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4dh8d"] Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.836500 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4dh8d"] Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.839933 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jrfsv"] Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.846289 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jrfsv"] Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.849381 4818 scope.go:117] "RemoveContainer" containerID="d53dac0a6d42f7ce63071a532981951505c9038d78d08c70cfe80d0d6bf73037" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.871351 4818 scope.go:117] "RemoveContainer" containerID="3e37f9cc5816052f7c6991762135723f0f86e6f484e32bd8e2870d46593d55bc" Dec 03 06:33:31 crc kubenswrapper[4818]: E1203 06:33:31.871972 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e37f9cc5816052f7c6991762135723f0f86e6f484e32bd8e2870d46593d55bc\": container with ID starting with 3e37f9cc5816052f7c6991762135723f0f86e6f484e32bd8e2870d46593d55bc not found: ID does not exist" containerID="3e37f9cc5816052f7c6991762135723f0f86e6f484e32bd8e2870d46593d55bc" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.872017 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e37f9cc5816052f7c6991762135723f0f86e6f484e32bd8e2870d46593d55bc"} err="failed to get container status \"3e37f9cc5816052f7c6991762135723f0f86e6f484e32bd8e2870d46593d55bc\": rpc error: code = NotFound desc = could not find container \"3e37f9cc5816052f7c6991762135723f0f86e6f484e32bd8e2870d46593d55bc\": container with ID starting with 3e37f9cc5816052f7c6991762135723f0f86e6f484e32bd8e2870d46593d55bc not found: ID does not exist" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.872046 4818 scope.go:117] "RemoveContainer" containerID="8f0fe5cc139e0b9e63edec1a51c00eac2621dccc305d594aa83e600ea5c4212c" Dec 03 06:33:31 crc kubenswrapper[4818]: E1203 06:33:31.872635 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f0fe5cc139e0b9e63edec1a51c00eac2621dccc305d594aa83e600ea5c4212c\": container with ID starting with 8f0fe5cc139e0b9e63edec1a51c00eac2621dccc305d594aa83e600ea5c4212c not found: ID does not exist" containerID="8f0fe5cc139e0b9e63edec1a51c00eac2621dccc305d594aa83e600ea5c4212c" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.872676 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f0fe5cc139e0b9e63edec1a51c00eac2621dccc305d594aa83e600ea5c4212c"} err="failed to get container status \"8f0fe5cc139e0b9e63edec1a51c00eac2621dccc305d594aa83e600ea5c4212c\": rpc error: code = NotFound desc = could not find container \"8f0fe5cc139e0b9e63edec1a51c00eac2621dccc305d594aa83e600ea5c4212c\": container with ID starting with 8f0fe5cc139e0b9e63edec1a51c00eac2621dccc305d594aa83e600ea5c4212c not found: ID does not exist" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.872704 4818 scope.go:117] "RemoveContainer" containerID="d53dac0a6d42f7ce63071a532981951505c9038d78d08c70cfe80d0d6bf73037" Dec 03 06:33:31 crc kubenswrapper[4818]: E1203 06:33:31.873767 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d53dac0a6d42f7ce63071a532981951505c9038d78d08c70cfe80d0d6bf73037\": container with ID starting with d53dac0a6d42f7ce63071a532981951505c9038d78d08c70cfe80d0d6bf73037 not found: ID does not exist" containerID="d53dac0a6d42f7ce63071a532981951505c9038d78d08c70cfe80d0d6bf73037" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.873812 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d53dac0a6d42f7ce63071a532981951505c9038d78d08c70cfe80d0d6bf73037"} err="failed to get container status \"d53dac0a6d42f7ce63071a532981951505c9038d78d08c70cfe80d0d6bf73037\": rpc error: code = NotFound desc = could not find container \"d53dac0a6d42f7ce63071a532981951505c9038d78d08c70cfe80d0d6bf73037\": container with ID starting with d53dac0a6d42f7ce63071a532981951505c9038d78d08c70cfe80d0d6bf73037 not found: ID does not exist" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.873868 4818 scope.go:117] "RemoveContainer" containerID="04caaf0e08ce8e688a22ab83ddfb7237545bbbb93f5e378ad7d0955d1d9446fd" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.894222 4818 scope.go:117] "RemoveContainer" containerID="beb806bf0b4a2470eb3fd8b472967f2c9d9a177842e7782b4cbb64047bdf3217" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.915296 4818 scope.go:117] "RemoveContainer" containerID="28c1098bef6a1e692c39f2a2050ee72cb8efdb87e3e4987f112d3233edf99904" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.932354 4818 scope.go:117] "RemoveContainer" containerID="04caaf0e08ce8e688a22ab83ddfb7237545bbbb93f5e378ad7d0955d1d9446fd" Dec 03 06:33:31 crc kubenswrapper[4818]: E1203 06:33:31.932861 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04caaf0e08ce8e688a22ab83ddfb7237545bbbb93f5e378ad7d0955d1d9446fd\": container with ID starting with 04caaf0e08ce8e688a22ab83ddfb7237545bbbb93f5e378ad7d0955d1d9446fd not found: ID does not exist" containerID="04caaf0e08ce8e688a22ab83ddfb7237545bbbb93f5e378ad7d0955d1d9446fd" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.932905 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04caaf0e08ce8e688a22ab83ddfb7237545bbbb93f5e378ad7d0955d1d9446fd"} err="failed to get container status \"04caaf0e08ce8e688a22ab83ddfb7237545bbbb93f5e378ad7d0955d1d9446fd\": rpc error: code = NotFound desc = could not find container \"04caaf0e08ce8e688a22ab83ddfb7237545bbbb93f5e378ad7d0955d1d9446fd\": container with ID starting with 04caaf0e08ce8e688a22ab83ddfb7237545bbbb93f5e378ad7d0955d1d9446fd not found: ID does not exist" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.932931 4818 scope.go:117] "RemoveContainer" containerID="beb806bf0b4a2470eb3fd8b472967f2c9d9a177842e7782b4cbb64047bdf3217" Dec 03 06:33:31 crc kubenswrapper[4818]: E1203 06:33:31.933261 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"beb806bf0b4a2470eb3fd8b472967f2c9d9a177842e7782b4cbb64047bdf3217\": container with ID starting with beb806bf0b4a2470eb3fd8b472967f2c9d9a177842e7782b4cbb64047bdf3217 not found: ID does not exist" containerID="beb806bf0b4a2470eb3fd8b472967f2c9d9a177842e7782b4cbb64047bdf3217" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.933287 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beb806bf0b4a2470eb3fd8b472967f2c9d9a177842e7782b4cbb64047bdf3217"} err="failed to get container status \"beb806bf0b4a2470eb3fd8b472967f2c9d9a177842e7782b4cbb64047bdf3217\": rpc error: code = NotFound desc = could not find container \"beb806bf0b4a2470eb3fd8b472967f2c9d9a177842e7782b4cbb64047bdf3217\": container with ID starting with beb806bf0b4a2470eb3fd8b472967f2c9d9a177842e7782b4cbb64047bdf3217 not found: ID does not exist" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.933305 4818 scope.go:117] "RemoveContainer" containerID="28c1098bef6a1e692c39f2a2050ee72cb8efdb87e3e4987f112d3233edf99904" Dec 03 06:33:31 crc kubenswrapper[4818]: E1203 06:33:31.933615 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28c1098bef6a1e692c39f2a2050ee72cb8efdb87e3e4987f112d3233edf99904\": container with ID starting with 28c1098bef6a1e692c39f2a2050ee72cb8efdb87e3e4987f112d3233edf99904 not found: ID does not exist" containerID="28c1098bef6a1e692c39f2a2050ee72cb8efdb87e3e4987f112d3233edf99904" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.933637 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28c1098bef6a1e692c39f2a2050ee72cb8efdb87e3e4987f112d3233edf99904"} err="failed to get container status \"28c1098bef6a1e692c39f2a2050ee72cb8efdb87e3e4987f112d3233edf99904\": rpc error: code = NotFound desc = could not find container \"28c1098bef6a1e692c39f2a2050ee72cb8efdb87e3e4987f112d3233edf99904\": container with ID starting with 28c1098bef6a1e692c39f2a2050ee72cb8efdb87e3e4987f112d3233edf99904 not found: ID does not exist" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.933652 4818 scope.go:117] "RemoveContainer" containerID="3395f8e86e03b477f14f091b8e9c92d7de78b57f50e6601bb62d2c7e1dbad671" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.946442 4818 scope.go:117] "RemoveContainer" containerID="dddd0b228babbc5d22cd3ceeca18a47a59bd31fc0b34de7805a71b039c0c2fcb" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.961027 4818 scope.go:117] "RemoveContainer" containerID="3395f8e86e03b477f14f091b8e9c92d7de78b57f50e6601bb62d2c7e1dbad671" Dec 03 06:33:31 crc kubenswrapper[4818]: E1203 06:33:31.961560 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3395f8e86e03b477f14f091b8e9c92d7de78b57f50e6601bb62d2c7e1dbad671\": container with ID starting with 3395f8e86e03b477f14f091b8e9c92d7de78b57f50e6601bb62d2c7e1dbad671 not found: ID does not exist" containerID="3395f8e86e03b477f14f091b8e9c92d7de78b57f50e6601bb62d2c7e1dbad671" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.961597 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3395f8e86e03b477f14f091b8e9c92d7de78b57f50e6601bb62d2c7e1dbad671"} err="failed to get container status \"3395f8e86e03b477f14f091b8e9c92d7de78b57f50e6601bb62d2c7e1dbad671\": rpc error: code = NotFound desc = could not find container \"3395f8e86e03b477f14f091b8e9c92d7de78b57f50e6601bb62d2c7e1dbad671\": container with ID starting with 3395f8e86e03b477f14f091b8e9c92d7de78b57f50e6601bb62d2c7e1dbad671 not found: ID does not exist" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.961624 4818 scope.go:117] "RemoveContainer" containerID="dddd0b228babbc5d22cd3ceeca18a47a59bd31fc0b34de7805a71b039c0c2fcb" Dec 03 06:33:31 crc kubenswrapper[4818]: E1203 06:33:31.961933 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dddd0b228babbc5d22cd3ceeca18a47a59bd31fc0b34de7805a71b039c0c2fcb\": container with ID starting with dddd0b228babbc5d22cd3ceeca18a47a59bd31fc0b34de7805a71b039c0c2fcb not found: ID does not exist" containerID="dddd0b228babbc5d22cd3ceeca18a47a59bd31fc0b34de7805a71b039c0c2fcb" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.961963 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dddd0b228babbc5d22cd3ceeca18a47a59bd31fc0b34de7805a71b039c0c2fcb"} err="failed to get container status \"dddd0b228babbc5d22cd3ceeca18a47a59bd31fc0b34de7805a71b039c0c2fcb\": rpc error: code = NotFound desc = could not find container \"dddd0b228babbc5d22cd3ceeca18a47a59bd31fc0b34de7805a71b039c0c2fcb\": container with ID starting with dddd0b228babbc5d22cd3ceeca18a47a59bd31fc0b34de7805a71b039c0c2fcb not found: ID does not exist" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.961983 4818 scope.go:117] "RemoveContainer" containerID="1ea4fe4551a71e496e406f333497c69daf15b0f43c41a115de3c6500e2520da6" Dec 03 06:33:31 crc kubenswrapper[4818]: I1203 06:33:31.975671 4818 scope.go:117] "RemoveContainer" containerID="196fff865cfdab8c5f02bfefaa00f1df7162323673ab198c96326c7a3f8c62db" Dec 03 06:33:32 crc kubenswrapper[4818]: I1203 06:33:32.007223 4818 scope.go:117] "RemoveContainer" containerID="6d01faa4d0822c30f72b3f4b18f389157c34e498fefe338180d930d5af9bbfd2" Dec 03 06:33:32 crc kubenswrapper[4818]: I1203 06:33:32.020341 4818 scope.go:117] "RemoveContainer" containerID="1ea4fe4551a71e496e406f333497c69daf15b0f43c41a115de3c6500e2520da6" Dec 03 06:33:32 crc kubenswrapper[4818]: E1203 06:33:32.021427 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ea4fe4551a71e496e406f333497c69daf15b0f43c41a115de3c6500e2520da6\": container with ID starting with 1ea4fe4551a71e496e406f333497c69daf15b0f43c41a115de3c6500e2520da6 not found: ID does not exist" containerID="1ea4fe4551a71e496e406f333497c69daf15b0f43c41a115de3c6500e2520da6" Dec 03 06:33:32 crc kubenswrapper[4818]: I1203 06:33:32.021468 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ea4fe4551a71e496e406f333497c69daf15b0f43c41a115de3c6500e2520da6"} err="failed to get container status \"1ea4fe4551a71e496e406f333497c69daf15b0f43c41a115de3c6500e2520da6\": rpc error: code = NotFound desc = could not find container \"1ea4fe4551a71e496e406f333497c69daf15b0f43c41a115de3c6500e2520da6\": container with ID starting with 1ea4fe4551a71e496e406f333497c69daf15b0f43c41a115de3c6500e2520da6 not found: ID does not exist" Dec 03 06:33:32 crc kubenswrapper[4818]: I1203 06:33:32.021492 4818 scope.go:117] "RemoveContainer" containerID="196fff865cfdab8c5f02bfefaa00f1df7162323673ab198c96326c7a3f8c62db" Dec 03 06:33:32 crc kubenswrapper[4818]: E1203 06:33:32.021849 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"196fff865cfdab8c5f02bfefaa00f1df7162323673ab198c96326c7a3f8c62db\": container with ID starting with 196fff865cfdab8c5f02bfefaa00f1df7162323673ab198c96326c7a3f8c62db not found: ID does not exist" containerID="196fff865cfdab8c5f02bfefaa00f1df7162323673ab198c96326c7a3f8c62db" Dec 03 06:33:32 crc kubenswrapper[4818]: I1203 06:33:32.021880 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"196fff865cfdab8c5f02bfefaa00f1df7162323673ab198c96326c7a3f8c62db"} err="failed to get container status \"196fff865cfdab8c5f02bfefaa00f1df7162323673ab198c96326c7a3f8c62db\": rpc error: code = NotFound desc = could not find container \"196fff865cfdab8c5f02bfefaa00f1df7162323673ab198c96326c7a3f8c62db\": container with ID starting with 196fff865cfdab8c5f02bfefaa00f1df7162323673ab198c96326c7a3f8c62db not found: ID does not exist" Dec 03 06:33:32 crc kubenswrapper[4818]: I1203 06:33:32.021897 4818 scope.go:117] "RemoveContainer" containerID="6d01faa4d0822c30f72b3f4b18f389157c34e498fefe338180d930d5af9bbfd2" Dec 03 06:33:32 crc kubenswrapper[4818]: E1203 06:33:32.023399 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d01faa4d0822c30f72b3f4b18f389157c34e498fefe338180d930d5af9bbfd2\": container with ID starting with 6d01faa4d0822c30f72b3f4b18f389157c34e498fefe338180d930d5af9bbfd2 not found: ID does not exist" containerID="6d01faa4d0822c30f72b3f4b18f389157c34e498fefe338180d930d5af9bbfd2" Dec 03 06:33:32 crc kubenswrapper[4818]: I1203 06:33:32.023432 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d01faa4d0822c30f72b3f4b18f389157c34e498fefe338180d930d5af9bbfd2"} err="failed to get container status \"6d01faa4d0822c30f72b3f4b18f389157c34e498fefe338180d930d5af9bbfd2\": rpc error: code = NotFound desc = could not find container \"6d01faa4d0822c30f72b3f4b18f389157c34e498fefe338180d930d5af9bbfd2\": container with ID starting with 6d01faa4d0822c30f72b3f4b18f389157c34e498fefe338180d930d5af9bbfd2 not found: ID does not exist" Dec 03 06:33:32 crc kubenswrapper[4818]: I1203 06:33:32.748153 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0301f892-ca1d-4f8b-a429-6d0ad96a50dc" path="/var/lib/kubelet/pods/0301f892-ca1d-4f8b-a429-6d0ad96a50dc/volumes" Dec 03 06:33:32 crc kubenswrapper[4818]: I1203 06:33:32.749016 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c4ffde8-459a-4634-b567-5f900d40ceec" path="/var/lib/kubelet/pods/0c4ffde8-459a-4634-b567-5f900d40ceec/volumes" Dec 03 06:33:32 crc kubenswrapper[4818]: I1203 06:33:32.750123 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53a413a7-34f8-4fb3-893b-c4fa9711017a" path="/var/lib/kubelet/pods/53a413a7-34f8-4fb3-893b-c4fa9711017a/volumes" Dec 03 06:33:32 crc kubenswrapper[4818]: I1203 06:33:32.751512 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75955bfe-ae6e-481d-84b5-3de486a86827" path="/var/lib/kubelet/pods/75955bfe-ae6e-481d-84b5-3de486a86827/volumes" Dec 03 06:33:32 crc kubenswrapper[4818]: I1203 06:33:32.752309 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91223086-535a-47c4-a843-3bb100a7da50" path="/var/lib/kubelet/pods/91223086-535a-47c4-a843-3bb100a7da50/volumes" Dec 03 06:33:32 crc kubenswrapper[4818]: I1203 06:33:32.764531 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-hl66m" Dec 03 06:33:43 crc kubenswrapper[4818]: I1203 06:33:43.302674 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:33:43 crc kubenswrapper[4818]: I1203 06:33:43.303405 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.843670 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hchtl"] Dec 03 06:33:52 crc kubenswrapper[4818]: E1203 06:33:52.844664 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75955bfe-ae6e-481d-84b5-3de486a86827" containerName="extract-content" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.844689 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="75955bfe-ae6e-481d-84b5-3de486a86827" containerName="extract-content" Dec 03 06:33:52 crc kubenswrapper[4818]: E1203 06:33:52.844708 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0301f892-ca1d-4f8b-a429-6d0ad96a50dc" containerName="registry-server" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.844737 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0301f892-ca1d-4f8b-a429-6d0ad96a50dc" containerName="registry-server" Dec 03 06:33:52 crc kubenswrapper[4818]: E1203 06:33:52.844757 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91223086-535a-47c4-a843-3bb100a7da50" containerName="extract-utilities" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.844770 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="91223086-535a-47c4-a843-3bb100a7da50" containerName="extract-utilities" Dec 03 06:33:52 crc kubenswrapper[4818]: E1203 06:33:52.844782 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53a413a7-34f8-4fb3-893b-c4fa9711017a" containerName="marketplace-operator" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.844794 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="53a413a7-34f8-4fb3-893b-c4fa9711017a" containerName="marketplace-operator" Dec 03 06:33:52 crc kubenswrapper[4818]: E1203 06:33:52.844811 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75955bfe-ae6e-481d-84b5-3de486a86827" containerName="registry-server" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.844851 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="75955bfe-ae6e-481d-84b5-3de486a86827" containerName="registry-server" Dec 03 06:33:52 crc kubenswrapper[4818]: E1203 06:33:52.844878 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c4ffde8-459a-4634-b567-5f900d40ceec" containerName="extract-content" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.844893 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c4ffde8-459a-4634-b567-5f900d40ceec" containerName="extract-content" Dec 03 06:33:52 crc kubenswrapper[4818]: E1203 06:33:52.844910 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91223086-535a-47c4-a843-3bb100a7da50" containerName="registry-server" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.844922 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="91223086-535a-47c4-a843-3bb100a7da50" containerName="registry-server" Dec 03 06:33:52 crc kubenswrapper[4818]: E1203 06:33:52.844941 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c4ffde8-459a-4634-b567-5f900d40ceec" containerName="extract-utilities" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.844953 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c4ffde8-459a-4634-b567-5f900d40ceec" containerName="extract-utilities" Dec 03 06:33:52 crc kubenswrapper[4818]: E1203 06:33:52.844973 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75955bfe-ae6e-481d-84b5-3de486a86827" containerName="extract-utilities" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.844985 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="75955bfe-ae6e-481d-84b5-3de486a86827" containerName="extract-utilities" Dec 03 06:33:52 crc kubenswrapper[4818]: E1203 06:33:52.845004 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53a413a7-34f8-4fb3-893b-c4fa9711017a" containerName="marketplace-operator" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.845016 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="53a413a7-34f8-4fb3-893b-c4fa9711017a" containerName="marketplace-operator" Dec 03 06:33:52 crc kubenswrapper[4818]: E1203 06:33:52.845032 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c4ffde8-459a-4634-b567-5f900d40ceec" containerName="registry-server" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.845045 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c4ffde8-459a-4634-b567-5f900d40ceec" containerName="registry-server" Dec 03 06:33:52 crc kubenswrapper[4818]: E1203 06:33:52.845066 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91223086-535a-47c4-a843-3bb100a7da50" containerName="extract-content" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.845078 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="91223086-535a-47c4-a843-3bb100a7da50" containerName="extract-content" Dec 03 06:33:52 crc kubenswrapper[4818]: E1203 06:33:52.845097 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0301f892-ca1d-4f8b-a429-6d0ad96a50dc" containerName="extract-utilities" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.845108 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0301f892-ca1d-4f8b-a429-6d0ad96a50dc" containerName="extract-utilities" Dec 03 06:33:52 crc kubenswrapper[4818]: E1203 06:33:52.845128 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0301f892-ca1d-4f8b-a429-6d0ad96a50dc" containerName="extract-content" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.845139 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0301f892-ca1d-4f8b-a429-6d0ad96a50dc" containerName="extract-content" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.845367 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="53a413a7-34f8-4fb3-893b-c4fa9711017a" containerName="marketplace-operator" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.845385 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="53a413a7-34f8-4fb3-893b-c4fa9711017a" containerName="marketplace-operator" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.845406 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c4ffde8-459a-4634-b567-5f900d40ceec" containerName="registry-server" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.845420 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="91223086-535a-47c4-a843-3bb100a7da50" containerName="registry-server" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.845439 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="0301f892-ca1d-4f8b-a429-6d0ad96a50dc" containerName="registry-server" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.845475 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="75955bfe-ae6e-481d-84b5-3de486a86827" containerName="registry-server" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.846916 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hchtl" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.850910 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.867055 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hchtl"] Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.970830 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/761a2fb4-d124-4db1-9924-0cd3d9546fa5-catalog-content\") pod \"redhat-operators-hchtl\" (UID: \"761a2fb4-d124-4db1-9924-0cd3d9546fa5\") " pod="openshift-marketplace/redhat-operators-hchtl" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.970931 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/761a2fb4-d124-4db1-9924-0cd3d9546fa5-utilities\") pod \"redhat-operators-hchtl\" (UID: \"761a2fb4-d124-4db1-9924-0cd3d9546fa5\") " pod="openshift-marketplace/redhat-operators-hchtl" Dec 03 06:33:52 crc kubenswrapper[4818]: I1203 06:33:52.970991 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9srs\" (UniqueName: \"kubernetes.io/projected/761a2fb4-d124-4db1-9924-0cd3d9546fa5-kube-api-access-w9srs\") pod \"redhat-operators-hchtl\" (UID: \"761a2fb4-d124-4db1-9924-0cd3d9546fa5\") " pod="openshift-marketplace/redhat-operators-hchtl" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.043708 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mwgrm"] Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.051444 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mwgrm" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.054355 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mwgrm"] Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.055672 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.072972 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/761a2fb4-d124-4db1-9924-0cd3d9546fa5-utilities\") pod \"redhat-operators-hchtl\" (UID: \"761a2fb4-d124-4db1-9924-0cd3d9546fa5\") " pod="openshift-marketplace/redhat-operators-hchtl" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.073018 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9srs\" (UniqueName: \"kubernetes.io/projected/761a2fb4-d124-4db1-9924-0cd3d9546fa5-kube-api-access-w9srs\") pod \"redhat-operators-hchtl\" (UID: \"761a2fb4-d124-4db1-9924-0cd3d9546fa5\") " pod="openshift-marketplace/redhat-operators-hchtl" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.073086 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/761a2fb4-d124-4db1-9924-0cd3d9546fa5-catalog-content\") pod \"redhat-operators-hchtl\" (UID: \"761a2fb4-d124-4db1-9924-0cd3d9546fa5\") " pod="openshift-marketplace/redhat-operators-hchtl" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.073501 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/761a2fb4-d124-4db1-9924-0cd3d9546fa5-catalog-content\") pod \"redhat-operators-hchtl\" (UID: \"761a2fb4-d124-4db1-9924-0cd3d9546fa5\") " pod="openshift-marketplace/redhat-operators-hchtl" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.073743 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/761a2fb4-d124-4db1-9924-0cd3d9546fa5-utilities\") pod \"redhat-operators-hchtl\" (UID: \"761a2fb4-d124-4db1-9924-0cd3d9546fa5\") " pod="openshift-marketplace/redhat-operators-hchtl" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.091526 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9srs\" (UniqueName: \"kubernetes.io/projected/761a2fb4-d124-4db1-9924-0cd3d9546fa5-kube-api-access-w9srs\") pod \"redhat-operators-hchtl\" (UID: \"761a2fb4-d124-4db1-9924-0cd3d9546fa5\") " pod="openshift-marketplace/redhat-operators-hchtl" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.174584 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57e20fce-c349-473f-8866-b194da09c27d-catalog-content\") pod \"community-operators-mwgrm\" (UID: \"57e20fce-c349-473f-8866-b194da09c27d\") " pod="openshift-marketplace/community-operators-mwgrm" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.175184 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq4z4\" (UniqueName: \"kubernetes.io/projected/57e20fce-c349-473f-8866-b194da09c27d-kube-api-access-zq4z4\") pod \"community-operators-mwgrm\" (UID: \"57e20fce-c349-473f-8866-b194da09c27d\") " pod="openshift-marketplace/community-operators-mwgrm" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.175248 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57e20fce-c349-473f-8866-b194da09c27d-utilities\") pod \"community-operators-mwgrm\" (UID: \"57e20fce-c349-473f-8866-b194da09c27d\") " pod="openshift-marketplace/community-operators-mwgrm" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.186069 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hchtl" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.276361 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57e20fce-c349-473f-8866-b194da09c27d-catalog-content\") pod \"community-operators-mwgrm\" (UID: \"57e20fce-c349-473f-8866-b194da09c27d\") " pod="openshift-marketplace/community-operators-mwgrm" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.276537 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq4z4\" (UniqueName: \"kubernetes.io/projected/57e20fce-c349-473f-8866-b194da09c27d-kube-api-access-zq4z4\") pod \"community-operators-mwgrm\" (UID: \"57e20fce-c349-473f-8866-b194da09c27d\") " pod="openshift-marketplace/community-operators-mwgrm" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.276637 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57e20fce-c349-473f-8866-b194da09c27d-utilities\") pod \"community-operators-mwgrm\" (UID: \"57e20fce-c349-473f-8866-b194da09c27d\") " pod="openshift-marketplace/community-operators-mwgrm" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.277191 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57e20fce-c349-473f-8866-b194da09c27d-catalog-content\") pod \"community-operators-mwgrm\" (UID: \"57e20fce-c349-473f-8866-b194da09c27d\") " pod="openshift-marketplace/community-operators-mwgrm" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.277249 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57e20fce-c349-473f-8866-b194da09c27d-utilities\") pod \"community-operators-mwgrm\" (UID: \"57e20fce-c349-473f-8866-b194da09c27d\") " pod="openshift-marketplace/community-operators-mwgrm" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.300789 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq4z4\" (UniqueName: \"kubernetes.io/projected/57e20fce-c349-473f-8866-b194da09c27d-kube-api-access-zq4z4\") pod \"community-operators-mwgrm\" (UID: \"57e20fce-c349-473f-8866-b194da09c27d\") " pod="openshift-marketplace/community-operators-mwgrm" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.389339 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mwgrm" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.633471 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hchtl"] Dec 03 06:33:53 crc kubenswrapper[4818]: W1203 06:33:53.641277 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod761a2fb4_d124_4db1_9924_0cd3d9546fa5.slice/crio-bbf0b06f96e910ca5df8a5866fb5a8b5806fb4397baaef9d3d36c0cf61e1210f WatchSource:0}: Error finding container bbf0b06f96e910ca5df8a5866fb5a8b5806fb4397baaef9d3d36c0cf61e1210f: Status 404 returned error can't find the container with id bbf0b06f96e910ca5df8a5866fb5a8b5806fb4397baaef9d3d36c0cf61e1210f Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.787127 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.787199 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.787247 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.787287 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.793521 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.795702 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mwgrm"] Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.798178 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.799369 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.799475 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.838712 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.897411 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwgrm" event={"ID":"57e20fce-c349-473f-8866-b194da09c27d","Type":"ContainerStarted","Data":"f4b652b6c084a48ff4d790896af57ad3bc2ffdef97bddd472376140d61b98952"} Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.899139 4818 generic.go:334] "Generic (PLEG): container finished" podID="761a2fb4-d124-4db1-9924-0cd3d9546fa5" containerID="d4244de0152a5461390683d1f943e998de84a7c74f03215a3ef12fd107747a25" exitCode=0 Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.899183 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hchtl" event={"ID":"761a2fb4-d124-4db1-9924-0cd3d9546fa5","Type":"ContainerDied","Data":"d4244de0152a5461390683d1f943e998de84a7c74f03215a3ef12fd107747a25"} Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.899213 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hchtl" event={"ID":"761a2fb4-d124-4db1-9924-0cd3d9546fa5","Type":"ContainerStarted","Data":"bbf0b06f96e910ca5df8a5866fb5a8b5806fb4397baaef9d3d36c0cf61e1210f"} Dec 03 06:33:53 crc kubenswrapper[4818]: I1203 06:33:53.941112 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 06:33:54 crc kubenswrapper[4818]: I1203 06:33:54.051028 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 06:33:54 crc kubenswrapper[4818]: W1203 06:33:54.149260 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-5d0d6abbbdfee6e23c936a5ab66a81ef4ef4bd1725d58c7ca67c1d486a1fc015 WatchSource:0}: Error finding container 5d0d6abbbdfee6e23c936a5ab66a81ef4ef4bd1725d58c7ca67c1d486a1fc015: Status 404 returned error can't find the container with id 5d0d6abbbdfee6e23c936a5ab66a81ef4ef4bd1725d58c7ca67c1d486a1fc015 Dec 03 06:33:54 crc kubenswrapper[4818]: I1203 06:33:54.907003 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"5d0d6abbbdfee6e23c936a5ab66a81ef4ef4bd1725d58c7ca67c1d486a1fc015"} Dec 03 06:33:54 crc kubenswrapper[4818]: I1203 06:33:54.909084 4818 generic.go:334] "Generic (PLEG): container finished" podID="57e20fce-c349-473f-8866-b194da09c27d" containerID="2aae74fe42431c099446f21e6a21ed6734d3888258544577eb1b4997855e210f" exitCode=0 Dec 03 06:33:54 crc kubenswrapper[4818]: I1203 06:33:54.909188 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwgrm" event={"ID":"57e20fce-c349-473f-8866-b194da09c27d","Type":"ContainerDied","Data":"2aae74fe42431c099446f21e6a21ed6734d3888258544577eb1b4997855e210f"} Dec 03 06:33:54 crc kubenswrapper[4818]: I1203 06:33:54.910944 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"dc090f96d9ecd58da288f53dc0f5761958de1089d04ae50d40ab9797ef93b10a"} Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.042123 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2kmpd"] Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.043357 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2kmpd" Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.046640 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.205678 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5jb6\" (UniqueName: \"kubernetes.io/projected/93370008-4229-4f0f-be59-8385406d7cd1-kube-api-access-q5jb6\") pod \"certified-operators-2kmpd\" (UID: \"93370008-4229-4f0f-be59-8385406d7cd1\") " pod="openshift-marketplace/certified-operators-2kmpd" Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.205750 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93370008-4229-4f0f-be59-8385406d7cd1-utilities\") pod \"certified-operators-2kmpd\" (UID: \"93370008-4229-4f0f-be59-8385406d7cd1\") " pod="openshift-marketplace/certified-operators-2kmpd" Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.205796 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93370008-4229-4f0f-be59-8385406d7cd1-catalog-content\") pod \"certified-operators-2kmpd\" (UID: \"93370008-4229-4f0f-be59-8385406d7cd1\") " pod="openshift-marketplace/certified-operators-2kmpd" Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.239810 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2kmpd"] Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.307196 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5jb6\" (UniqueName: \"kubernetes.io/projected/93370008-4229-4f0f-be59-8385406d7cd1-kube-api-access-q5jb6\") pod \"certified-operators-2kmpd\" (UID: \"93370008-4229-4f0f-be59-8385406d7cd1\") " pod="openshift-marketplace/certified-operators-2kmpd" Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.307238 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93370008-4229-4f0f-be59-8385406d7cd1-utilities\") pod \"certified-operators-2kmpd\" (UID: \"93370008-4229-4f0f-be59-8385406d7cd1\") " pod="openshift-marketplace/certified-operators-2kmpd" Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.307271 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93370008-4229-4f0f-be59-8385406d7cd1-catalog-content\") pod \"certified-operators-2kmpd\" (UID: \"93370008-4229-4f0f-be59-8385406d7cd1\") " pod="openshift-marketplace/certified-operators-2kmpd" Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.307842 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93370008-4229-4f0f-be59-8385406d7cd1-catalog-content\") pod \"certified-operators-2kmpd\" (UID: \"93370008-4229-4f0f-be59-8385406d7cd1\") " pod="openshift-marketplace/certified-operators-2kmpd" Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.307977 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93370008-4229-4f0f-be59-8385406d7cd1-utilities\") pod \"certified-operators-2kmpd\" (UID: \"93370008-4229-4f0f-be59-8385406d7cd1\") " pod="openshift-marketplace/certified-operators-2kmpd" Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.327580 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5jb6\" (UniqueName: \"kubernetes.io/projected/93370008-4229-4f0f-be59-8385406d7cd1-kube-api-access-q5jb6\") pod \"certified-operators-2kmpd\" (UID: \"93370008-4229-4f0f-be59-8385406d7cd1\") " pod="openshift-marketplace/certified-operators-2kmpd" Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.371290 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2kmpd" Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.786933 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2kmpd"] Dec 03 06:33:55 crc kubenswrapper[4818]: W1203 06:33:55.790909 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93370008_4229_4f0f_be59_8385406d7cd1.slice/crio-fd558b1dfbe7f47b53b822d7ba0e810462f45bf53e2011b01f44ba84860b4767 WatchSource:0}: Error finding container fd558b1dfbe7f47b53b822d7ba0e810462f45bf53e2011b01f44ba84860b4767: Status 404 returned error can't find the container with id fd558b1dfbe7f47b53b822d7ba0e810462f45bf53e2011b01f44ba84860b4767 Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.838807 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dm6xl"] Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.839733 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dm6xl" Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.848642 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.850869 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dm6xl"] Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.917622 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"be5b3af8ea4bf41c40d7b74606a61f89bf6830a08fc23fc9c82faf537c1a8f5e"} Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.918788 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2kmpd" event={"ID":"93370008-4229-4f0f-be59-8385406d7cd1","Type":"ContainerStarted","Data":"fd558b1dfbe7f47b53b822d7ba0e810462f45bf53e2011b01f44ba84860b4767"} Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.921254 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e7b4b545f29781b9190e11226cca6eb1ed3965587f17ee236d45803490644094"} Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.921280 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"9d72b9089a83856cf8ec03074d943d5e3329e2e7b20b30985fa923474514c72f"} Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.923090 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"6e343558b67144876184f45d6c48c27de0868a3f748d3795916c43af72777539"} Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.923592 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:33:55 crc kubenswrapper[4818]: I1203 06:33:55.925841 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hchtl" event={"ID":"761a2fb4-d124-4db1-9924-0cd3d9546fa5","Type":"ContainerStarted","Data":"8a6ed09b2af2f6cf06be8f9d4e57530c904fbdcf13ab78711719b7d179a21f82"} Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.020086 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2a86a74-f04c-49ed-a900-33a852b794a3-catalog-content\") pod \"redhat-marketplace-dm6xl\" (UID: \"b2a86a74-f04c-49ed-a900-33a852b794a3\") " pod="openshift-marketplace/redhat-marketplace-dm6xl" Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.020156 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2a86a74-f04c-49ed-a900-33a852b794a3-utilities\") pod \"redhat-marketplace-dm6xl\" (UID: \"b2a86a74-f04c-49ed-a900-33a852b794a3\") " pod="openshift-marketplace/redhat-marketplace-dm6xl" Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.020320 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c64wj\" (UniqueName: \"kubernetes.io/projected/b2a86a74-f04c-49ed-a900-33a852b794a3-kube-api-access-c64wj\") pod \"redhat-marketplace-dm6xl\" (UID: \"b2a86a74-f04c-49ed-a900-33a852b794a3\") " pod="openshift-marketplace/redhat-marketplace-dm6xl" Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.121930 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2a86a74-f04c-49ed-a900-33a852b794a3-utilities\") pod \"redhat-marketplace-dm6xl\" (UID: \"b2a86a74-f04c-49ed-a900-33a852b794a3\") " pod="openshift-marketplace/redhat-marketplace-dm6xl" Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.121998 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c64wj\" (UniqueName: \"kubernetes.io/projected/b2a86a74-f04c-49ed-a900-33a852b794a3-kube-api-access-c64wj\") pod \"redhat-marketplace-dm6xl\" (UID: \"b2a86a74-f04c-49ed-a900-33a852b794a3\") " pod="openshift-marketplace/redhat-marketplace-dm6xl" Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.122027 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2a86a74-f04c-49ed-a900-33a852b794a3-catalog-content\") pod \"redhat-marketplace-dm6xl\" (UID: \"b2a86a74-f04c-49ed-a900-33a852b794a3\") " pod="openshift-marketplace/redhat-marketplace-dm6xl" Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.122418 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2a86a74-f04c-49ed-a900-33a852b794a3-utilities\") pod \"redhat-marketplace-dm6xl\" (UID: \"b2a86a74-f04c-49ed-a900-33a852b794a3\") " pod="openshift-marketplace/redhat-marketplace-dm6xl" Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.122455 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2a86a74-f04c-49ed-a900-33a852b794a3-catalog-content\") pod \"redhat-marketplace-dm6xl\" (UID: \"b2a86a74-f04c-49ed-a900-33a852b794a3\") " pod="openshift-marketplace/redhat-marketplace-dm6xl" Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.141308 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c64wj\" (UniqueName: \"kubernetes.io/projected/b2a86a74-f04c-49ed-a900-33a852b794a3-kube-api-access-c64wj\") pod \"redhat-marketplace-dm6xl\" (UID: \"b2a86a74-f04c-49ed-a900-33a852b794a3\") " pod="openshift-marketplace/redhat-marketplace-dm6xl" Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.297209 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dm6xl" Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.511483 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dm6xl"] Dec 03 06:33:56 crc kubenswrapper[4818]: W1203 06:33:56.516149 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2a86a74_f04c_49ed_a900_33a852b794a3.slice/crio-9546744fa8a10a34c33b1e09a7eb6783310994fa07cd04838b6bc33ec7e06817 WatchSource:0}: Error finding container 9546744fa8a10a34c33b1e09a7eb6783310994fa07cd04838b6bc33ec7e06817: Status 404 returned error can't find the container with id 9546744fa8a10a34c33b1e09a7eb6783310994fa07cd04838b6bc33ec7e06817 Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.934436 4818 generic.go:334] "Generic (PLEG): container finished" podID="761a2fb4-d124-4db1-9924-0cd3d9546fa5" containerID="8a6ed09b2af2f6cf06be8f9d4e57530c904fbdcf13ab78711719b7d179a21f82" exitCode=0 Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.934751 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hchtl" event={"ID":"761a2fb4-d124-4db1-9924-0cd3d9546fa5","Type":"ContainerDied","Data":"8a6ed09b2af2f6cf06be8f9d4e57530c904fbdcf13ab78711719b7d179a21f82"} Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.939141 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm6xl" event={"ID":"b2a86a74-f04c-49ed-a900-33a852b794a3","Type":"ContainerDied","Data":"6daad1e46c076dbc8ed1ff17703cc1ea3da1c74c2c980927cf5a4c1f02644a13"} Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.939996 4818 generic.go:334] "Generic (PLEG): container finished" podID="b2a86a74-f04c-49ed-a900-33a852b794a3" containerID="6daad1e46c076dbc8ed1ff17703cc1ea3da1c74c2c980927cf5a4c1f02644a13" exitCode=0 Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.940130 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm6xl" event={"ID":"b2a86a74-f04c-49ed-a900-33a852b794a3","Type":"ContainerStarted","Data":"9546744fa8a10a34c33b1e09a7eb6783310994fa07cd04838b6bc33ec7e06817"} Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.942596 4818 generic.go:334] "Generic (PLEG): container finished" podID="93370008-4229-4f0f-be59-8385406d7cd1" containerID="d5909db946ba1064f007a66379359c1217a56b1298a01ee1ee19de0d2ef0a278" exitCode=0 Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.942642 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2kmpd" event={"ID":"93370008-4229-4f0f-be59-8385406d7cd1","Type":"ContainerDied","Data":"d5909db946ba1064f007a66379359c1217a56b1298a01ee1ee19de0d2ef0a278"} Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.948576 4818 generic.go:334] "Generic (PLEG): container finished" podID="57e20fce-c349-473f-8866-b194da09c27d" containerID="a997770edc0889bbee6d1b674b13edd67f1631820f46dd673b3e53e9359a087b" exitCode=0 Dec 03 06:33:56 crc kubenswrapper[4818]: I1203 06:33:56.948671 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwgrm" event={"ID":"57e20fce-c349-473f-8866-b194da09c27d","Type":"ContainerDied","Data":"a997770edc0889bbee6d1b674b13edd67f1631820f46dd673b3e53e9359a087b"} Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.335694 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bg9rn"] Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.336456 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.403885 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bg9rn"] Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.539153 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6krbt\" (UniqueName: \"kubernetes.io/projected/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-kube-api-access-6krbt\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.539198 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-bound-sa-token\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.539222 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.539270 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-registry-tls\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.539295 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-trusted-ca\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.539335 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-registry-certificates\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.539371 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.539427 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.565612 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.640371 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-registry-tls\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.640425 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-trusted-ca\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.640453 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-registry-certificates\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.640510 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.640540 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6krbt\" (UniqueName: \"kubernetes.io/projected/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-kube-api-access-6krbt\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.640558 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-bound-sa-token\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.640576 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.641276 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.642015 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-trusted-ca\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.642167 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-registry-certificates\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.651276 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-registry-tls\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.654286 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.656339 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-bound-sa-token\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.656708 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6krbt\" (UniqueName: \"kubernetes.io/projected/91634bb1-e7bc-43e4-8212-fd528d6a2ae5-kube-api-access-6krbt\") pod \"image-registry-66df7c8f76-bg9rn\" (UID: \"91634bb1-e7bc-43e4-8212-fd528d6a2ae5\") " pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.659588 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.957396 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwgrm" event={"ID":"57e20fce-c349-473f-8866-b194da09c27d","Type":"ContainerStarted","Data":"afde14fa7c5a8f50dd12f8f9c872f17a9f54cfda6fe8090cd71eb8fb5744ec4a"} Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.961335 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hchtl" event={"ID":"761a2fb4-d124-4db1-9924-0cd3d9546fa5","Type":"ContainerStarted","Data":"875364e16cbab99bcddb4a399af532d8d0eb469a56fdc0bb39a86aa97e190241"} Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.963709 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm6xl" event={"ID":"b2a86a74-f04c-49ed-a900-33a852b794a3","Type":"ContainerStarted","Data":"64ed5a0a8153c3d81a8aeeb9834469e7d70cb1008eee46777d98f11b3eb3decf"} Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.966191 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2kmpd" event={"ID":"93370008-4229-4f0f-be59-8385406d7cd1","Type":"ContainerStarted","Data":"e739400f09ea1254260d7a8ad03550ef792b7b6298857ccdbfa0c2067b570231"} Dec 03 06:33:57 crc kubenswrapper[4818]: I1203 06:33:57.976265 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mwgrm" podStartSLOduration=2.773082104 podStartE2EDuration="4.976246237s" podCreationTimestamp="2025-12-03 06:33:53 +0000 UTC" firstStartedPulling="2025-12-03 06:33:55.23640616 +0000 UTC m=+392.928014912" lastFinishedPulling="2025-12-03 06:33:57.439570293 +0000 UTC m=+395.131179045" observedRunningTime="2025-12-03 06:33:57.975435947 +0000 UTC m=+395.667044709" watchObservedRunningTime="2025-12-03 06:33:57.976246237 +0000 UTC m=+395.667854989" Dec 03 06:33:58 crc kubenswrapper[4818]: I1203 06:33:58.011753 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bg9rn"] Dec 03 06:33:58 crc kubenswrapper[4818]: I1203 06:33:58.014495 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hchtl" podStartSLOduration=2.371413701 podStartE2EDuration="6.014473627s" podCreationTimestamp="2025-12-03 06:33:52 +0000 UTC" firstStartedPulling="2025-12-03 06:33:53.902676899 +0000 UTC m=+391.594285691" lastFinishedPulling="2025-12-03 06:33:57.545736835 +0000 UTC m=+395.237345617" observedRunningTime="2025-12-03 06:33:58.009148586 +0000 UTC m=+395.700757338" watchObservedRunningTime="2025-12-03 06:33:58.014473627 +0000 UTC m=+395.706082379" Dec 03 06:33:58 crc kubenswrapper[4818]: I1203 06:33:58.973086 4818 generic.go:334] "Generic (PLEG): container finished" podID="b2a86a74-f04c-49ed-a900-33a852b794a3" containerID="64ed5a0a8153c3d81a8aeeb9834469e7d70cb1008eee46777d98f11b3eb3decf" exitCode=0 Dec 03 06:33:58 crc kubenswrapper[4818]: I1203 06:33:58.973158 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm6xl" event={"ID":"b2a86a74-f04c-49ed-a900-33a852b794a3","Type":"ContainerDied","Data":"64ed5a0a8153c3d81a8aeeb9834469e7d70cb1008eee46777d98f11b3eb3decf"} Dec 03 06:33:58 crc kubenswrapper[4818]: I1203 06:33:58.975196 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" event={"ID":"91634bb1-e7bc-43e4-8212-fd528d6a2ae5","Type":"ContainerStarted","Data":"4f1efeace4a36065f8c7b786450d587f583b7983db97c253fbe54eb650ed1e57"} Dec 03 06:33:58 crc kubenswrapper[4818]: I1203 06:33:58.975228 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" event={"ID":"91634bb1-e7bc-43e4-8212-fd528d6a2ae5","Type":"ContainerStarted","Data":"315fc8a3f9e1d894184b679b6d3615f5e7dbaefe983819cbbc4c2e7d7b079ad8"} Dec 03 06:33:58 crc kubenswrapper[4818]: I1203 06:33:58.975404 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:33:58 crc kubenswrapper[4818]: I1203 06:33:58.979748 4818 generic.go:334] "Generic (PLEG): container finished" podID="93370008-4229-4f0f-be59-8385406d7cd1" containerID="e739400f09ea1254260d7a8ad03550ef792b7b6298857ccdbfa0c2067b570231" exitCode=0 Dec 03 06:33:58 crc kubenswrapper[4818]: I1203 06:33:58.981351 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2kmpd" event={"ID":"93370008-4229-4f0f-be59-8385406d7cd1","Type":"ContainerDied","Data":"e739400f09ea1254260d7a8ad03550ef792b7b6298857ccdbfa0c2067b570231"} Dec 03 06:33:59 crc kubenswrapper[4818]: I1203 06:33:59.987722 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2kmpd" event={"ID":"93370008-4229-4f0f-be59-8385406d7cd1","Type":"ContainerStarted","Data":"0ec4a4c6ecc2afee246fb45562939d564fffc4ea06349e5d160032aad4339883"} Dec 03 06:33:59 crc kubenswrapper[4818]: I1203 06:33:59.990965 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm6xl" event={"ID":"b2a86a74-f04c-49ed-a900-33a852b794a3","Type":"ContainerStarted","Data":"4f695d8e8cf74e2d38e47c5d9409a4497797df9339471d4f9c969bbb047029eb"} Dec 03 06:34:00 crc kubenswrapper[4818]: I1203 06:34:00.008606 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2kmpd" podStartSLOduration=2.532804636 podStartE2EDuration="5.008590567s" podCreationTimestamp="2025-12-03 06:33:55 +0000 UTC" firstStartedPulling="2025-12-03 06:33:56.94437487 +0000 UTC m=+394.635983632" lastFinishedPulling="2025-12-03 06:33:59.420160811 +0000 UTC m=+397.111769563" observedRunningTime="2025-12-03 06:34:00.005260095 +0000 UTC m=+397.696868847" watchObservedRunningTime="2025-12-03 06:34:00.008590567 +0000 UTC m=+397.700199309" Dec 03 06:34:00 crc kubenswrapper[4818]: I1203 06:34:00.008917 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" podStartSLOduration=3.008910745 podStartE2EDuration="3.008910745s" podCreationTimestamp="2025-12-03 06:33:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:33:59.048357413 +0000 UTC m=+396.739966205" watchObservedRunningTime="2025-12-03 06:34:00.008910745 +0000 UTC m=+397.700519497" Dec 03 06:34:00 crc kubenswrapper[4818]: I1203 06:34:00.030746 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dm6xl" podStartSLOduration=2.423236181 podStartE2EDuration="5.030727832s" podCreationTimestamp="2025-12-03 06:33:55 +0000 UTC" firstStartedPulling="2025-12-03 06:33:56.941192122 +0000 UTC m=+394.632800874" lastFinishedPulling="2025-12-03 06:33:59.548683773 +0000 UTC m=+397.240292525" observedRunningTime="2025-12-03 06:34:00.025665097 +0000 UTC m=+397.717273859" watchObservedRunningTime="2025-12-03 06:34:00.030727832 +0000 UTC m=+397.722336594" Dec 03 06:34:03 crc kubenswrapper[4818]: I1203 06:34:03.186765 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hchtl" Dec 03 06:34:03 crc kubenswrapper[4818]: I1203 06:34:03.187883 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hchtl" Dec 03 06:34:03 crc kubenswrapper[4818]: I1203 06:34:03.389843 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mwgrm" Dec 03 06:34:03 crc kubenswrapper[4818]: I1203 06:34:03.392963 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mwgrm" Dec 03 06:34:03 crc kubenswrapper[4818]: I1203 06:34:03.440240 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mwgrm" Dec 03 06:34:04 crc kubenswrapper[4818]: I1203 06:34:04.062243 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mwgrm" Dec 03 06:34:04 crc kubenswrapper[4818]: I1203 06:34:04.242974 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hchtl" podUID="761a2fb4-d124-4db1-9924-0cd3d9546fa5" containerName="registry-server" probeResult="failure" output=< Dec 03 06:34:04 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Dec 03 06:34:04 crc kubenswrapper[4818]: > Dec 03 06:34:05 crc kubenswrapper[4818]: I1203 06:34:05.371871 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2kmpd" Dec 03 06:34:05 crc kubenswrapper[4818]: I1203 06:34:05.371925 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2kmpd" Dec 03 06:34:05 crc kubenswrapper[4818]: I1203 06:34:05.430305 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2kmpd" Dec 03 06:34:06 crc kubenswrapper[4818]: I1203 06:34:06.113027 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2kmpd" Dec 03 06:34:06 crc kubenswrapper[4818]: I1203 06:34:06.298089 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dm6xl" Dec 03 06:34:06 crc kubenswrapper[4818]: I1203 06:34:06.298149 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dm6xl" Dec 03 06:34:06 crc kubenswrapper[4818]: I1203 06:34:06.347499 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dm6xl" Dec 03 06:34:07 crc kubenswrapper[4818]: I1203 06:34:07.104167 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dm6xl" Dec 03 06:34:13 crc kubenswrapper[4818]: I1203 06:34:13.257557 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hchtl" Dec 03 06:34:13 crc kubenswrapper[4818]: I1203 06:34:13.303153 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:34:13 crc kubenswrapper[4818]: I1203 06:34:13.303273 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:34:13 crc kubenswrapper[4818]: I1203 06:34:13.303459 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:34:13 crc kubenswrapper[4818]: I1203 06:34:13.305152 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1f4c468b483e867be2291edf0efa4884963d305fb4fe62f714e10f36adc23d54"} pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 06:34:13 crc kubenswrapper[4818]: I1203 06:34:13.305285 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" containerID="cri-o://1f4c468b483e867be2291edf0efa4884963d305fb4fe62f714e10f36adc23d54" gracePeriod=600 Dec 03 06:34:13 crc kubenswrapper[4818]: I1203 06:34:13.334968 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hchtl" Dec 03 06:34:15 crc kubenswrapper[4818]: I1203 06:34:15.096653 4818 generic.go:334] "Generic (PLEG): container finished" podID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerID="1f4c468b483e867be2291edf0efa4884963d305fb4fe62f714e10f36adc23d54" exitCode=0 Dec 03 06:34:15 crc kubenswrapper[4818]: I1203 06:34:15.096746 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerDied","Data":"1f4c468b483e867be2291edf0efa4884963d305fb4fe62f714e10f36adc23d54"} Dec 03 06:34:15 crc kubenswrapper[4818]: I1203 06:34:15.097148 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"29214294c382d4bed0dc9a820848900ea80c550461b1b157cc6bdd05a78333e5"} Dec 03 06:34:15 crc kubenswrapper[4818]: I1203 06:34:15.097185 4818 scope.go:117] "RemoveContainer" containerID="a9869627bb82825beecc4b9986f953389d3f4557d4f472a01e131e28b015401f" Dec 03 06:34:17 crc kubenswrapper[4818]: I1203 06:34:17.665479 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-bg9rn" Dec 03 06:34:17 crc kubenswrapper[4818]: I1203 06:34:17.720167 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zg4cb"] Dec 03 06:34:33 crc kubenswrapper[4818]: I1203 06:34:33.844603 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 06:34:42 crc kubenswrapper[4818]: I1203 06:34:42.762733 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" podUID="535884b5-264b-4471-962c-51579eeacab0" containerName="registry" containerID="cri-o://78bb80e7ab7e9c0fdbe26cb87cb807c0970253b99e1e0d90089eea4780144efb" gracePeriod=30 Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.189496 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.270983 4818 generic.go:334] "Generic (PLEG): container finished" podID="535884b5-264b-4471-962c-51579eeacab0" containerID="78bb80e7ab7e9c0fdbe26cb87cb807c0970253b99e1e0d90089eea4780144efb" exitCode=0 Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.271046 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.271036 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" event={"ID":"535884b5-264b-4471-962c-51579eeacab0","Type":"ContainerDied","Data":"78bb80e7ab7e9c0fdbe26cb87cb807c0970253b99e1e0d90089eea4780144efb"} Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.271578 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-zg4cb" event={"ID":"535884b5-264b-4471-962c-51579eeacab0","Type":"ContainerDied","Data":"0335d8d890192bf536f57525605b2579f9e0e5b041909ae284dc7bd718feffe1"} Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.271612 4818 scope.go:117] "RemoveContainer" containerID="78bb80e7ab7e9c0fdbe26cb87cb807c0970253b99e1e0d90089eea4780144efb" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.290479 4818 scope.go:117] "RemoveContainer" containerID="78bb80e7ab7e9c0fdbe26cb87cb807c0970253b99e1e0d90089eea4780144efb" Dec 03 06:34:43 crc kubenswrapper[4818]: E1203 06:34:43.291113 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78bb80e7ab7e9c0fdbe26cb87cb807c0970253b99e1e0d90089eea4780144efb\": container with ID starting with 78bb80e7ab7e9c0fdbe26cb87cb807c0970253b99e1e0d90089eea4780144efb not found: ID does not exist" containerID="78bb80e7ab7e9c0fdbe26cb87cb807c0970253b99e1e0d90089eea4780144efb" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.291161 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78bb80e7ab7e9c0fdbe26cb87cb807c0970253b99e1e0d90089eea4780144efb"} err="failed to get container status \"78bb80e7ab7e9c0fdbe26cb87cb807c0970253b99e1e0d90089eea4780144efb\": rpc error: code = NotFound desc = could not find container \"78bb80e7ab7e9c0fdbe26cb87cb807c0970253b99e1e0d90089eea4780144efb\": container with ID starting with 78bb80e7ab7e9c0fdbe26cb87cb807c0970253b99e1e0d90089eea4780144efb not found: ID does not exist" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.390043 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnlc6\" (UniqueName: \"kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-kube-api-access-gnlc6\") pod \"535884b5-264b-4471-962c-51579eeacab0\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.390089 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/535884b5-264b-4471-962c-51579eeacab0-installation-pull-secrets\") pod \"535884b5-264b-4471-962c-51579eeacab0\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.390294 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"535884b5-264b-4471-962c-51579eeacab0\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.390327 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/535884b5-264b-4471-962c-51579eeacab0-trusted-ca\") pod \"535884b5-264b-4471-962c-51579eeacab0\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.390364 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/535884b5-264b-4471-962c-51579eeacab0-ca-trust-extracted\") pod \"535884b5-264b-4471-962c-51579eeacab0\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.390386 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/535884b5-264b-4471-962c-51579eeacab0-registry-certificates\") pod \"535884b5-264b-4471-962c-51579eeacab0\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.390466 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-registry-tls\") pod \"535884b5-264b-4471-962c-51579eeacab0\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.390512 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-bound-sa-token\") pod \"535884b5-264b-4471-962c-51579eeacab0\" (UID: \"535884b5-264b-4471-962c-51579eeacab0\") " Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.392460 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/535884b5-264b-4471-962c-51579eeacab0-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "535884b5-264b-4471-962c-51579eeacab0" (UID: "535884b5-264b-4471-962c-51579eeacab0"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.393021 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/535884b5-264b-4471-962c-51579eeacab0-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "535884b5-264b-4471-962c-51579eeacab0" (UID: "535884b5-264b-4471-962c-51579eeacab0"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.396964 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-kube-api-access-gnlc6" (OuterVolumeSpecName: "kube-api-access-gnlc6") pod "535884b5-264b-4471-962c-51579eeacab0" (UID: "535884b5-264b-4471-962c-51579eeacab0"). InnerVolumeSpecName "kube-api-access-gnlc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.398222 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/535884b5-264b-4471-962c-51579eeacab0-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "535884b5-264b-4471-962c-51579eeacab0" (UID: "535884b5-264b-4471-962c-51579eeacab0"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.400388 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "535884b5-264b-4471-962c-51579eeacab0" (UID: "535884b5-264b-4471-962c-51579eeacab0"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.401194 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "535884b5-264b-4471-962c-51579eeacab0" (UID: "535884b5-264b-4471-962c-51579eeacab0"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.426526 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/535884b5-264b-4471-962c-51579eeacab0-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "535884b5-264b-4471-962c-51579eeacab0" (UID: "535884b5-264b-4471-962c-51579eeacab0"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.426888 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "535884b5-264b-4471-962c-51579eeacab0" (UID: "535884b5-264b-4471-962c-51579eeacab0"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.491403 4818 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.491450 4818 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.491467 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnlc6\" (UniqueName: \"kubernetes.io/projected/535884b5-264b-4471-962c-51579eeacab0-kube-api-access-gnlc6\") on node \"crc\" DevicePath \"\"" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.491486 4818 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/535884b5-264b-4471-962c-51579eeacab0-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.491500 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/535884b5-264b-4471-962c-51579eeacab0-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.491514 4818 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/535884b5-264b-4471-962c-51579eeacab0-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.491529 4818 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/535884b5-264b-4471-962c-51579eeacab0-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.630247 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zg4cb"] Dec 03 06:34:43 crc kubenswrapper[4818]: I1203 06:34:43.644199 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-zg4cb"] Dec 03 06:34:44 crc kubenswrapper[4818]: I1203 06:34:44.747895 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="535884b5-264b-4471-962c-51579eeacab0" path="/var/lib/kubelet/pods/535884b5-264b-4471-962c-51579eeacab0/volumes" Dec 03 06:36:43 crc kubenswrapper[4818]: I1203 06:36:43.302303 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:36:43 crc kubenswrapper[4818]: I1203 06:36:43.302958 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:37:13 crc kubenswrapper[4818]: I1203 06:37:13.302488 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:37:13 crc kubenswrapper[4818]: I1203 06:37:13.303222 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:37:43 crc kubenswrapper[4818]: I1203 06:37:43.302791 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:37:43 crc kubenswrapper[4818]: I1203 06:37:43.303742 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:37:43 crc kubenswrapper[4818]: I1203 06:37:43.303807 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:37:43 crc kubenswrapper[4818]: I1203 06:37:43.304652 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"29214294c382d4bed0dc9a820848900ea80c550461b1b157cc6bdd05a78333e5"} pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 06:37:43 crc kubenswrapper[4818]: I1203 06:37:43.304739 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" containerID="cri-o://29214294c382d4bed0dc9a820848900ea80c550461b1b157cc6bdd05a78333e5" gracePeriod=600 Dec 03 06:37:43 crc kubenswrapper[4818]: I1203 06:37:43.467342 4818 generic.go:334] "Generic (PLEG): container finished" podID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerID="29214294c382d4bed0dc9a820848900ea80c550461b1b157cc6bdd05a78333e5" exitCode=0 Dec 03 06:37:43 crc kubenswrapper[4818]: I1203 06:37:43.467425 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerDied","Data":"29214294c382d4bed0dc9a820848900ea80c550461b1b157cc6bdd05a78333e5"} Dec 03 06:37:43 crc kubenswrapper[4818]: I1203 06:37:43.467653 4818 scope.go:117] "RemoveContainer" containerID="1f4c468b483e867be2291edf0efa4884963d305fb4fe62f714e10f36adc23d54" Dec 03 06:37:44 crc kubenswrapper[4818]: I1203 06:37:44.478230 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"d9ea0215376ae671f51a5593609153a4204668620744066ed3e3215741cd9e5f"} Dec 03 06:38:51 crc kubenswrapper[4818]: I1203 06:38:51.843556 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-cpcd5"] Dec 03 06:38:51 crc kubenswrapper[4818]: E1203 06:38:51.844443 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="535884b5-264b-4471-962c-51579eeacab0" containerName="registry" Dec 03 06:38:51 crc kubenswrapper[4818]: I1203 06:38:51.844461 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="535884b5-264b-4471-962c-51579eeacab0" containerName="registry" Dec 03 06:38:51 crc kubenswrapper[4818]: I1203 06:38:51.844570 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="535884b5-264b-4471-962c-51579eeacab0" containerName="registry" Dec 03 06:38:51 crc kubenswrapper[4818]: I1203 06:38:51.844969 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-cpcd5" Dec 03 06:38:51 crc kubenswrapper[4818]: I1203 06:38:51.849311 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xl5nj"] Dec 03 06:38:51 crc kubenswrapper[4818]: I1203 06:38:51.849900 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-xl5nj" Dec 03 06:38:51 crc kubenswrapper[4818]: I1203 06:38:51.856615 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 03 06:38:51 crc kubenswrapper[4818]: I1203 06:38:51.856980 4818 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-njzd4" Dec 03 06:38:51 crc kubenswrapper[4818]: I1203 06:38:51.857250 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-cpcd5"] Dec 03 06:38:51 crc kubenswrapper[4818]: I1203 06:38:51.857619 4818 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-h6p8t" Dec 03 06:38:51 crc kubenswrapper[4818]: I1203 06:38:51.858618 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 03 06:38:51 crc kubenswrapper[4818]: I1203 06:38:51.869736 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xl5nj"] Dec 03 06:38:51 crc kubenswrapper[4818]: I1203 06:38:51.884360 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-tcn75"] Dec 03 06:38:51 crc kubenswrapper[4818]: I1203 06:38:51.885047 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-tcn75" Dec 03 06:38:51 crc kubenswrapper[4818]: I1203 06:38:51.887518 4818 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-8h852" Dec 03 06:38:51 crc kubenswrapper[4818]: I1203 06:38:51.903063 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-tcn75"] Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.008222 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpbzk\" (UniqueName: \"kubernetes.io/projected/8f6981d3-9529-498d-935e-f99f533faa9e-kube-api-access-hpbzk\") pod \"cert-manager-webhook-5655c58dd6-tcn75\" (UID: \"8f6981d3-9529-498d-935e-f99f533faa9e\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-tcn75" Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.008306 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmhhk\" (UniqueName: \"kubernetes.io/projected/c9bd0ac7-83ba-4f70-9458-0d8b88618456-kube-api-access-zmhhk\") pod \"cert-manager-cainjector-7f985d654d-cpcd5\" (UID: \"c9bd0ac7-83ba-4f70-9458-0d8b88618456\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-cpcd5" Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.008358 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djh82\" (UniqueName: \"kubernetes.io/projected/47f0d000-e043-4e9e-b300-988f03341b42-kube-api-access-djh82\") pod \"cert-manager-5b446d88c5-xl5nj\" (UID: \"47f0d000-e043-4e9e-b300-988f03341b42\") " pod="cert-manager/cert-manager-5b446d88c5-xl5nj" Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.109165 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmhhk\" (UniqueName: \"kubernetes.io/projected/c9bd0ac7-83ba-4f70-9458-0d8b88618456-kube-api-access-zmhhk\") pod \"cert-manager-cainjector-7f985d654d-cpcd5\" (UID: \"c9bd0ac7-83ba-4f70-9458-0d8b88618456\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-cpcd5" Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.109285 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djh82\" (UniqueName: \"kubernetes.io/projected/47f0d000-e043-4e9e-b300-988f03341b42-kube-api-access-djh82\") pod \"cert-manager-5b446d88c5-xl5nj\" (UID: \"47f0d000-e043-4e9e-b300-988f03341b42\") " pod="cert-manager/cert-manager-5b446d88c5-xl5nj" Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.109371 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpbzk\" (UniqueName: \"kubernetes.io/projected/8f6981d3-9529-498d-935e-f99f533faa9e-kube-api-access-hpbzk\") pod \"cert-manager-webhook-5655c58dd6-tcn75\" (UID: \"8f6981d3-9529-498d-935e-f99f533faa9e\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-tcn75" Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.130527 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpbzk\" (UniqueName: \"kubernetes.io/projected/8f6981d3-9529-498d-935e-f99f533faa9e-kube-api-access-hpbzk\") pod \"cert-manager-webhook-5655c58dd6-tcn75\" (UID: \"8f6981d3-9529-498d-935e-f99f533faa9e\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-tcn75" Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.135416 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djh82\" (UniqueName: \"kubernetes.io/projected/47f0d000-e043-4e9e-b300-988f03341b42-kube-api-access-djh82\") pod \"cert-manager-5b446d88c5-xl5nj\" (UID: \"47f0d000-e043-4e9e-b300-988f03341b42\") " pod="cert-manager/cert-manager-5b446d88c5-xl5nj" Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.136288 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmhhk\" (UniqueName: \"kubernetes.io/projected/c9bd0ac7-83ba-4f70-9458-0d8b88618456-kube-api-access-zmhhk\") pod \"cert-manager-cainjector-7f985d654d-cpcd5\" (UID: \"c9bd0ac7-83ba-4f70-9458-0d8b88618456\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-cpcd5" Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.189598 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-xl5nj" Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.189615 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-cpcd5" Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.206748 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-tcn75" Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.452261 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-tcn75"] Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.465953 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.492182 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-cpcd5"] Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.750733 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-xl5nj"] Dec 03 06:38:52 crc kubenswrapper[4818]: W1203 06:38:52.754228 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47f0d000_e043_4e9e_b300_988f03341b42.slice/crio-b57807a61e8ca34d0251ff20c2f40ae68232da9a019514c7c9f8811002c1089b WatchSource:0}: Error finding container b57807a61e8ca34d0251ff20c2f40ae68232da9a019514c7c9f8811002c1089b: Status 404 returned error can't find the container with id b57807a61e8ca34d0251ff20c2f40ae68232da9a019514c7c9f8811002c1089b Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.922303 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-xl5nj" event={"ID":"47f0d000-e043-4e9e-b300-988f03341b42","Type":"ContainerStarted","Data":"b57807a61e8ca34d0251ff20c2f40ae68232da9a019514c7c9f8811002c1089b"} Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.923479 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-cpcd5" event={"ID":"c9bd0ac7-83ba-4f70-9458-0d8b88618456","Type":"ContainerStarted","Data":"36d2ad2c2228c0ea6a29e36ea61a87b1362e1a208af8b859b01844ea0331917d"} Dec 03 06:38:52 crc kubenswrapper[4818]: I1203 06:38:52.924835 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-tcn75" event={"ID":"8f6981d3-9529-498d-935e-f99f533faa9e","Type":"ContainerStarted","Data":"633e111be868f443db368dc5794efe53109f2cc779da9547ca43c0e7fc583dac"} Dec 03 06:38:55 crc kubenswrapper[4818]: I1203 06:38:55.945205 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-tcn75" event={"ID":"8f6981d3-9529-498d-935e-f99f533faa9e","Type":"ContainerStarted","Data":"bc1cf17e49a3bd2506869d6e1190dc8604698611a73c6d2cdb08d4792d78284c"} Dec 03 06:38:55 crc kubenswrapper[4818]: I1203 06:38:55.945760 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-tcn75" Dec 03 06:38:55 crc kubenswrapper[4818]: I1203 06:38:55.947111 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-xl5nj" event={"ID":"47f0d000-e043-4e9e-b300-988f03341b42","Type":"ContainerStarted","Data":"17df7970c455c411236526c3a6e42051623ee2a2c4ba4f5d9283d192eca6c63c"} Dec 03 06:38:55 crc kubenswrapper[4818]: I1203 06:38:55.948483 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-cpcd5" event={"ID":"c9bd0ac7-83ba-4f70-9458-0d8b88618456","Type":"ContainerStarted","Data":"6d976d7b2988c7b127453786b008f7fdb1f8213f2a06a19e670fb88c1d6b2d89"} Dec 03 06:38:55 crc kubenswrapper[4818]: I1203 06:38:55.970649 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-tcn75" podStartSLOduration=1.845807789 podStartE2EDuration="4.970625746s" podCreationTimestamp="2025-12-03 06:38:51 +0000 UTC" firstStartedPulling="2025-12-03 06:38:52.465610148 +0000 UTC m=+690.157218900" lastFinishedPulling="2025-12-03 06:38:55.590428065 +0000 UTC m=+693.282036857" observedRunningTime="2025-12-03 06:38:55.966193086 +0000 UTC m=+693.657801838" watchObservedRunningTime="2025-12-03 06:38:55.970625746 +0000 UTC m=+693.662234508" Dec 03 06:38:56 crc kubenswrapper[4818]: I1203 06:38:56.003285 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-xl5nj" podStartSLOduration=2.167068949 podStartE2EDuration="5.003262454s" podCreationTimestamp="2025-12-03 06:38:51 +0000 UTC" firstStartedPulling="2025-12-03 06:38:52.756023494 +0000 UTC m=+690.447632256" lastFinishedPulling="2025-12-03 06:38:55.592216999 +0000 UTC m=+693.283825761" observedRunningTime="2025-12-03 06:38:56.000069965 +0000 UTC m=+693.691678737" watchObservedRunningTime="2025-12-03 06:38:56.003262454 +0000 UTC m=+693.694871206" Dec 03 06:38:56 crc kubenswrapper[4818]: I1203 06:38:56.065645 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-cpcd5" podStartSLOduration=1.966789986 podStartE2EDuration="5.06562971s" podCreationTimestamp="2025-12-03 06:38:51 +0000 UTC" firstStartedPulling="2025-12-03 06:38:52.501886636 +0000 UTC m=+690.193495388" lastFinishedPulling="2025-12-03 06:38:55.60072636 +0000 UTC m=+693.292335112" observedRunningTime="2025-12-03 06:38:56.052680649 +0000 UTC m=+693.744289391" watchObservedRunningTime="2025-12-03 06:38:56.06562971 +0000 UTC m=+693.757238462" Dec 03 06:39:02 crc kubenswrapper[4818]: I1203 06:39:02.210871 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-tcn75" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.067735 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ktm8k"] Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.069243 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovn-controller" containerID="cri-o://03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0" gracePeriod=30 Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.069959 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="sbdb" containerID="cri-o://837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7" gracePeriod=30 Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.070041 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="nbdb" containerID="cri-o://6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc" gracePeriod=30 Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.070109 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="northd" containerID="cri-o://4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643" gracePeriod=30 Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.070305 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a" gracePeriod=30 Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.070379 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="kube-rbac-proxy-node" containerID="cri-o://b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0" gracePeriod=30 Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.070438 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovn-acl-logging" containerID="cri-o://7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61" gracePeriod=30 Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.148741 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovnkube-controller" containerID="cri-o://98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc" gracePeriod=30 Dec 03 06:39:21 crc kubenswrapper[4818]: E1203 06:39:21.477343 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b25b836_b639_4111_bcea_af7cc791ea32.slice/crio-conmon-3860840e364d23f4d54acf2e87b4f131ee6d2bbd80175613d39d96953478088c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1e30fe6_0b94_4c95_b941_c0623729e123.slice/crio-conmon-98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1e30fe6_0b94_4c95_b941_c0623729e123.slice/crio-conmon-5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1e30fe6_0b94_4c95_b941_c0623729e123.slice/crio-conmon-b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0.scope\": RecentStats: unable to find data in memory cache]" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.801800 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovnkube-controller/3.log" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.805776 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovn-acl-logging/0.log" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.806912 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovn-controller/0.log" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.808222 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895083 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gzr42"] Dec 03 06:39:21 crc kubenswrapper[4818]: E1203 06:39:21.895391 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovnkube-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895411 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovnkube-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: E1203 06:39:21.895426 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovnkube-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895439 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovnkube-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: E1203 06:39:21.895456 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="sbdb" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895469 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="sbdb" Dec 03 06:39:21 crc kubenswrapper[4818]: E1203 06:39:21.895493 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="kubecfg-setup" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895504 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="kubecfg-setup" Dec 03 06:39:21 crc kubenswrapper[4818]: E1203 06:39:21.895520 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895532 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 06:39:21 crc kubenswrapper[4818]: E1203 06:39:21.895547 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="nbdb" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895559 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="nbdb" Dec 03 06:39:21 crc kubenswrapper[4818]: E1203 06:39:21.895584 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="kube-rbac-proxy-node" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895596 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="kube-rbac-proxy-node" Dec 03 06:39:21 crc kubenswrapper[4818]: E1203 06:39:21.895615 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovn-acl-logging" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895628 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovn-acl-logging" Dec 03 06:39:21 crc kubenswrapper[4818]: E1203 06:39:21.895643 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="northd" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895655 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="northd" Dec 03 06:39:21 crc kubenswrapper[4818]: E1203 06:39:21.895672 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovn-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895685 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovn-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: E1203 06:39:21.895698 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovnkube-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895709 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovnkube-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895888 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="kube-rbac-proxy-node" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895905 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895924 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="sbdb" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895940 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovnkube-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895956 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="nbdb" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895975 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="northd" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.895990 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovnkube-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.896005 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovn-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.896020 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovn-acl-logging" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.896036 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovnkube-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.896050 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovnkube-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: E1203 06:39:21.896215 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovnkube-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.896229 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovnkube-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: E1203 06:39:21.896246 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovnkube-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.896258 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovnkube-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.896456 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerName="ovnkube-controller" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.899580 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.916455 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-log-socket\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.916545 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a1e30fe6-0b94-4c95-b941-c0623729e123-ovn-node-metrics-cert\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.916595 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-var-lib-openvswitch\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.916649 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-var-lib-cni-networks-ovn-kubernetes\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.916699 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-run-netns\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.916793 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-systemd-units\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.916872 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-log-socket" (OuterVolumeSpecName: "log-socket") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.916910 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.916897 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-cni-bin\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.916978 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.916977 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917010 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917023 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917068 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-systemd\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917177 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-kubelet\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917219 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-etc-openvswitch\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917252 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917276 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-node-log\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917312 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917317 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-cni-netd\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917363 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917376 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-run-ovn-kubernetes\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917370 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-node-log" (OuterVolumeSpecName: "node-log") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917408 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917430 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrtqq\" (UniqueName: \"kubernetes.io/projected/a1e30fe6-0b94-4c95-b941-c0623729e123-kube-api-access-hrtqq\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917529 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-ovnkube-config\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917566 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-ovn\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917623 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-openvswitch\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917656 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-slash\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917686 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-ovnkube-script-lib\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917732 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-env-overrides\") pod \"a1e30fe6-0b94-4c95-b941-c0623729e123\" (UID: \"a1e30fe6-0b94-4c95-b941-c0623729e123\") " Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.917997 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.918379 4818 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.918409 4818 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.918429 4818 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-node-log\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.918447 4818 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.918464 4818 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.918483 4818 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.918501 4818 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-log-socket\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.918517 4818 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.918535 4818 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.918553 4818 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.918573 4818 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.918591 4818 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.918642 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-slash" (OuterVolumeSpecName: "host-slash") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.918990 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.919370 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.919393 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.919799 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.929529 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1e30fe6-0b94-4c95-b941-c0623729e123-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.929849 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1e30fe6-0b94-4c95-b941-c0623729e123-kube-api-access-hrtqq" (OuterVolumeSpecName: "kube-api-access-hrtqq") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "kube-api-access-hrtqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:39:21 crc kubenswrapper[4818]: I1203 06:39:21.957394 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "a1e30fe6-0b94-4c95-b941-c0623729e123" (UID: "a1e30fe6-0b94-4c95-b941-c0623729e123"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.019955 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-cni-netd\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020006 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d550a052-662c-48b1-9ef2-087a18ad2b4d-ovnkube-config\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020027 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d550a052-662c-48b1-9ef2-087a18ad2b4d-ovnkube-script-lib\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020046 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-run-ovn-kubernetes\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020061 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-node-log\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020087 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-kubelet\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020111 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-etc-openvswitch\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020287 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-var-lib-openvswitch\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020349 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-cni-bin\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020378 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-run-netns\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020400 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9dvd\" (UniqueName: \"kubernetes.io/projected/d550a052-662c-48b1-9ef2-087a18ad2b4d-kube-api-access-s9dvd\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020458 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-log-socket\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020500 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-slash\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020526 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-systemd-units\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020552 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-run-systemd\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020593 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020703 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d550a052-662c-48b1-9ef2-087a18ad2b4d-env-overrides\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020769 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d550a052-662c-48b1-9ef2-087a18ad2b4d-ovn-node-metrics-cert\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020794 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-run-openvswitch\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020833 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-run-ovn\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020942 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrtqq\" (UniqueName: \"kubernetes.io/projected/a1e30fe6-0b94-4c95-b941-c0623729e123-kube-api-access-hrtqq\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020975 4818 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020985 4818 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.020993 4818 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-host-slash\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.021001 4818 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.021009 4818 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a1e30fe6-0b94-4c95-b941-c0623729e123-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.021017 4818 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a1e30fe6-0b94-4c95-b941-c0623729e123-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.021029 4818 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a1e30fe6-0b94-4c95-b941-c0623729e123-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122336 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-kubelet\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122398 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-etc-openvswitch\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122430 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-var-lib-openvswitch\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122459 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-cni-bin\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122464 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-kubelet\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122483 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-run-netns\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122509 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-etc-openvswitch\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122533 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-run-netns\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122526 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9dvd\" (UniqueName: \"kubernetes.io/projected/d550a052-662c-48b1-9ef2-087a18ad2b4d-kube-api-access-s9dvd\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122603 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-cni-bin\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122590 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-var-lib-openvswitch\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122644 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-log-socket\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122683 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-slash\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122719 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-systemd-units\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122737 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-log-socket\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122771 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-systemd-units\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122780 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-run-systemd\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122754 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-run-systemd\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122748 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-slash\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122881 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122939 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d550a052-662c-48b1-9ef2-087a18ad2b4d-env-overrides\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.122997 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.123004 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d550a052-662c-48b1-9ef2-087a18ad2b4d-ovn-node-metrics-cert\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.123042 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-run-openvswitch\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.123058 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-run-ovn\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.123107 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-run-openvswitch\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.123116 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-cni-netd\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.123136 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-cni-netd\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.123158 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d550a052-662c-48b1-9ef2-087a18ad2b4d-ovnkube-config\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.123201 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d550a052-662c-48b1-9ef2-087a18ad2b4d-ovnkube-script-lib\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.123232 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-run-ovn-kubernetes\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.123161 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-run-ovn\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.123295 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-node-log\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.123263 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-node-log\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.123395 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d550a052-662c-48b1-9ef2-087a18ad2b4d-host-run-ovn-kubernetes\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.124009 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d550a052-662c-48b1-9ef2-087a18ad2b4d-ovnkube-config\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.124196 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d550a052-662c-48b1-9ef2-087a18ad2b4d-ovnkube-script-lib\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.124474 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d550a052-662c-48b1-9ef2-087a18ad2b4d-env-overrides\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.127115 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d550a052-662c-48b1-9ef2-087a18ad2b4d-ovn-node-metrics-cert\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.148254 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ctpzs_7b25b836-b639-4111-bcea-af7cc791ea32/kube-multus/2.log" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.148769 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ctpzs_7b25b836-b639-4111-bcea-af7cc791ea32/kube-multus/1.log" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.148809 4818 generic.go:334] "Generic (PLEG): container finished" podID="7b25b836-b639-4111-bcea-af7cc791ea32" containerID="3860840e364d23f4d54acf2e87b4f131ee6d2bbd80175613d39d96953478088c" exitCode=2 Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.148888 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ctpzs" event={"ID":"7b25b836-b639-4111-bcea-af7cc791ea32","Type":"ContainerDied","Data":"3860840e364d23f4d54acf2e87b4f131ee6d2bbd80175613d39d96953478088c"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.148935 4818 scope.go:117] "RemoveContainer" containerID="8b92f0662c82eef1fb0023107da93fead648febd6e3880b77ad264ced57de7be" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.149743 4818 scope.go:117] "RemoveContainer" containerID="3860840e364d23f4d54acf2e87b4f131ee6d2bbd80175613d39d96953478088c" Dec 03 06:39:22 crc kubenswrapper[4818]: E1203 06:39:22.150186 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-ctpzs_openshift-multus(7b25b836-b639-4111-bcea-af7cc791ea32)\"" pod="openshift-multus/multus-ctpzs" podUID="7b25b836-b639-4111-bcea-af7cc791ea32" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.151561 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovnkube-controller/3.log" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.155609 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovn-acl-logging/0.log" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.156807 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ktm8k_a1e30fe6-0b94-4c95-b941-c0623729e123/ovn-controller/0.log" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.157409 4818 generic.go:334] "Generic (PLEG): container finished" podID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerID="98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc" exitCode=0 Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.157464 4818 generic.go:334] "Generic (PLEG): container finished" podID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerID="837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7" exitCode=0 Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.157486 4818 generic.go:334] "Generic (PLEG): container finished" podID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerID="6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc" exitCode=0 Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.157505 4818 generic.go:334] "Generic (PLEG): container finished" podID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerID="4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643" exitCode=0 Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.157525 4818 generic.go:334] "Generic (PLEG): container finished" podID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerID="5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a" exitCode=0 Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.157526 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerDied","Data":"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.157590 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.157617 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerDied","Data":"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.157801 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerDied","Data":"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.157864 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerDied","Data":"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.157542 4818 generic.go:334] "Generic (PLEG): container finished" podID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerID="b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0" exitCode=0 Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.157944 4818 generic.go:334] "Generic (PLEG): container finished" podID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerID="7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61" exitCode=143 Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.157988 4818 generic.go:334] "Generic (PLEG): container finished" podID="a1e30fe6-0b94-4c95-b941-c0623729e123" containerID="03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0" exitCode=143 Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.157886 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerDied","Data":"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158067 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerDied","Data":"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158107 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158132 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158148 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158162 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158176 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158190 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158204 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158218 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158232 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158246 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158268 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerDied","Data":"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158292 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158309 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158324 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158338 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158352 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158367 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158381 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158394 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158408 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158422 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158442 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerDied","Data":"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158467 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158486 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158500 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158514 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158528 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158542 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158557 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158572 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158586 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158601 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158621 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ktm8k" event={"ID":"a1e30fe6-0b94-4c95-b941-c0623729e123","Type":"ContainerDied","Data":"cc3f0301bc894b4e2819c9d25e37cdeca35b35f3e7b4c11fb332c9c8653378e9"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158644 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158661 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158676 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158690 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158704 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158718 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158731 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158745 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158759 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.158772 4818 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16"} Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.168077 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9dvd\" (UniqueName: \"kubernetes.io/projected/d550a052-662c-48b1-9ef2-087a18ad2b4d-kube-api-access-s9dvd\") pod \"ovnkube-node-gzr42\" (UID: \"d550a052-662c-48b1-9ef2-087a18ad2b4d\") " pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.185927 4818 scope.go:117] "RemoveContainer" containerID="98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.226902 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ktm8k"] Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.229682 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.230742 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ktm8k"] Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.247662 4818 scope.go:117] "RemoveContainer" containerID="0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.273035 4818 scope.go:117] "RemoveContainer" containerID="837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.320263 4818 scope.go:117] "RemoveContainer" containerID="6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.357594 4818 scope.go:117] "RemoveContainer" containerID="4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.375883 4818 scope.go:117] "RemoveContainer" containerID="5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.410426 4818 scope.go:117] "RemoveContainer" containerID="b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.430400 4818 scope.go:117] "RemoveContainer" containerID="7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.455932 4818 scope.go:117] "RemoveContainer" containerID="03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.513320 4818 scope.go:117] "RemoveContainer" containerID="16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.530099 4818 scope.go:117] "RemoveContainer" containerID="98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc" Dec 03 06:39:22 crc kubenswrapper[4818]: E1203 06:39:22.533209 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc\": container with ID starting with 98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc not found: ID does not exist" containerID="98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.533247 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc"} err="failed to get container status \"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc\": rpc error: code = NotFound desc = could not find container \"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc\": container with ID starting with 98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.533271 4818 scope.go:117] "RemoveContainer" containerID="0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024" Dec 03 06:39:22 crc kubenswrapper[4818]: E1203 06:39:22.533692 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024\": container with ID starting with 0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024 not found: ID does not exist" containerID="0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.533775 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024"} err="failed to get container status \"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024\": rpc error: code = NotFound desc = could not find container \"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024\": container with ID starting with 0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.533898 4818 scope.go:117] "RemoveContainer" containerID="837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7" Dec 03 06:39:22 crc kubenswrapper[4818]: E1203 06:39:22.534382 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\": container with ID starting with 837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7 not found: ID does not exist" containerID="837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.534412 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7"} err="failed to get container status \"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\": rpc error: code = NotFound desc = could not find container \"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\": container with ID starting with 837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.534432 4818 scope.go:117] "RemoveContainer" containerID="6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc" Dec 03 06:39:22 crc kubenswrapper[4818]: E1203 06:39:22.534763 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\": container with ID starting with 6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc not found: ID does not exist" containerID="6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.534807 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc"} err="failed to get container status \"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\": rpc error: code = NotFound desc = could not find container \"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\": container with ID starting with 6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.534861 4818 scope.go:117] "RemoveContainer" containerID="4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643" Dec 03 06:39:22 crc kubenswrapper[4818]: E1203 06:39:22.535330 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\": container with ID starting with 4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643 not found: ID does not exist" containerID="4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.535364 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643"} err="failed to get container status \"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\": rpc error: code = NotFound desc = could not find container \"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\": container with ID starting with 4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.535391 4818 scope.go:117] "RemoveContainer" containerID="5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a" Dec 03 06:39:22 crc kubenswrapper[4818]: E1203 06:39:22.535727 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\": container with ID starting with 5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a not found: ID does not exist" containerID="5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.535764 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a"} err="failed to get container status \"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\": rpc error: code = NotFound desc = could not find container \"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\": container with ID starting with 5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.535787 4818 scope.go:117] "RemoveContainer" containerID="b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0" Dec 03 06:39:22 crc kubenswrapper[4818]: E1203 06:39:22.536138 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\": container with ID starting with b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0 not found: ID does not exist" containerID="b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.536182 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0"} err="failed to get container status \"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\": rpc error: code = NotFound desc = could not find container \"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\": container with ID starting with b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.536206 4818 scope.go:117] "RemoveContainer" containerID="7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61" Dec 03 06:39:22 crc kubenswrapper[4818]: E1203 06:39:22.536798 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\": container with ID starting with 7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61 not found: ID does not exist" containerID="7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.536841 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61"} err="failed to get container status \"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\": rpc error: code = NotFound desc = could not find container \"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\": container with ID starting with 7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.536856 4818 scope.go:117] "RemoveContainer" containerID="03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0" Dec 03 06:39:22 crc kubenswrapper[4818]: E1203 06:39:22.537163 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\": container with ID starting with 03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0 not found: ID does not exist" containerID="03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.537193 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0"} err="failed to get container status \"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\": rpc error: code = NotFound desc = could not find container \"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\": container with ID starting with 03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.537213 4818 scope.go:117] "RemoveContainer" containerID="16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16" Dec 03 06:39:22 crc kubenswrapper[4818]: E1203 06:39:22.537602 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\": container with ID starting with 16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16 not found: ID does not exist" containerID="16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.538219 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16"} err="failed to get container status \"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\": rpc error: code = NotFound desc = could not find container \"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\": container with ID starting with 16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.538248 4818 scope.go:117] "RemoveContainer" containerID="98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.538676 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc"} err="failed to get container status \"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc\": rpc error: code = NotFound desc = could not find container \"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc\": container with ID starting with 98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.538700 4818 scope.go:117] "RemoveContainer" containerID="0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.539032 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024"} err="failed to get container status \"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024\": rpc error: code = NotFound desc = could not find container \"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024\": container with ID starting with 0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.539069 4818 scope.go:117] "RemoveContainer" containerID="837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.539653 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7"} err="failed to get container status \"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\": rpc error: code = NotFound desc = could not find container \"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\": container with ID starting with 837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.539679 4818 scope.go:117] "RemoveContainer" containerID="6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.539983 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc"} err="failed to get container status \"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\": rpc error: code = NotFound desc = could not find container \"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\": container with ID starting with 6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.540018 4818 scope.go:117] "RemoveContainer" containerID="4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.540484 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643"} err="failed to get container status \"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\": rpc error: code = NotFound desc = could not find container \"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\": container with ID starting with 4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.540507 4818 scope.go:117] "RemoveContainer" containerID="5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.540896 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a"} err="failed to get container status \"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\": rpc error: code = NotFound desc = could not find container \"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\": container with ID starting with 5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.540930 4818 scope.go:117] "RemoveContainer" containerID="b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.541236 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0"} err="failed to get container status \"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\": rpc error: code = NotFound desc = could not find container \"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\": container with ID starting with b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.541268 4818 scope.go:117] "RemoveContainer" containerID="7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.541653 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61"} err="failed to get container status \"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\": rpc error: code = NotFound desc = could not find container \"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\": container with ID starting with 7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.541695 4818 scope.go:117] "RemoveContainer" containerID="03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.542043 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0"} err="failed to get container status \"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\": rpc error: code = NotFound desc = could not find container \"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\": container with ID starting with 03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.542083 4818 scope.go:117] "RemoveContainer" containerID="16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.542530 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16"} err="failed to get container status \"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\": rpc error: code = NotFound desc = could not find container \"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\": container with ID starting with 16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.542554 4818 scope.go:117] "RemoveContainer" containerID="98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.543089 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc"} err="failed to get container status \"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc\": rpc error: code = NotFound desc = could not find container \"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc\": container with ID starting with 98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.543247 4818 scope.go:117] "RemoveContainer" containerID="0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.543692 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024"} err="failed to get container status \"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024\": rpc error: code = NotFound desc = could not find container \"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024\": container with ID starting with 0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.543729 4818 scope.go:117] "RemoveContainer" containerID="837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.544077 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7"} err="failed to get container status \"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\": rpc error: code = NotFound desc = could not find container \"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\": container with ID starting with 837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.544145 4818 scope.go:117] "RemoveContainer" containerID="6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.544701 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc"} err="failed to get container status \"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\": rpc error: code = NotFound desc = could not find container \"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\": container with ID starting with 6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.544735 4818 scope.go:117] "RemoveContainer" containerID="4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.545213 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643"} err="failed to get container status \"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\": rpc error: code = NotFound desc = could not find container \"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\": container with ID starting with 4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.545243 4818 scope.go:117] "RemoveContainer" containerID="5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.545701 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a"} err="failed to get container status \"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\": rpc error: code = NotFound desc = could not find container \"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\": container with ID starting with 5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.545792 4818 scope.go:117] "RemoveContainer" containerID="b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.546108 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0"} err="failed to get container status \"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\": rpc error: code = NotFound desc = could not find container \"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\": container with ID starting with b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.546142 4818 scope.go:117] "RemoveContainer" containerID="7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.546401 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61"} err="failed to get container status \"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\": rpc error: code = NotFound desc = could not find container \"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\": container with ID starting with 7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.546424 4818 scope.go:117] "RemoveContainer" containerID="03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.546643 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0"} err="failed to get container status \"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\": rpc error: code = NotFound desc = could not find container \"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\": container with ID starting with 03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.546660 4818 scope.go:117] "RemoveContainer" containerID="16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.546912 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16"} err="failed to get container status \"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\": rpc error: code = NotFound desc = could not find container \"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\": container with ID starting with 16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.546956 4818 scope.go:117] "RemoveContainer" containerID="98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.547443 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc"} err="failed to get container status \"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc\": rpc error: code = NotFound desc = could not find container \"98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc\": container with ID starting with 98c9439dba5e8ea318b3a136e8ec1b3125a468aad364408ceb1687e2e9b74fcc not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.547464 4818 scope.go:117] "RemoveContainer" containerID="0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.547689 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024"} err="failed to get container status \"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024\": rpc error: code = NotFound desc = could not find container \"0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024\": container with ID starting with 0fabe012fa0de7026556a9f2a0c361bed165f480afffb91511e36dc80b35f024 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.547713 4818 scope.go:117] "RemoveContainer" containerID="837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.547950 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7"} err="failed to get container status \"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\": rpc error: code = NotFound desc = could not find container \"837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7\": container with ID starting with 837b57200be63769c264eb7147c99b8af71b7bd4f76b9ff2b959e74dc14513c7 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.547987 4818 scope.go:117] "RemoveContainer" containerID="6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.548286 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc"} err="failed to get container status \"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\": rpc error: code = NotFound desc = could not find container \"6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc\": container with ID starting with 6caad2aee10a5b7d8662c102981c373566043856b0d238bf5dc518eb5f5e0cdc not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.548308 4818 scope.go:117] "RemoveContainer" containerID="4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.548537 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643"} err="failed to get container status \"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\": rpc error: code = NotFound desc = could not find container \"4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643\": container with ID starting with 4e5b0275e3d56bdf56d14ef5da6d8562224dbbd1930679d1e60f11552f30e643 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.548560 4818 scope.go:117] "RemoveContainer" containerID="5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.548903 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a"} err="failed to get container status \"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\": rpc error: code = NotFound desc = could not find container \"5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a\": container with ID starting with 5f5862fc9f2008a356aaf21dfa200869f54f680bdd68f3197243c9391e3da12a not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.548954 4818 scope.go:117] "RemoveContainer" containerID="b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.549196 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0"} err="failed to get container status \"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\": rpc error: code = NotFound desc = could not find container \"b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0\": container with ID starting with b7bf04885a45388efea2d43f779033644194d01c7a6a23cb1807eaa80ce437d0 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.549217 4818 scope.go:117] "RemoveContainer" containerID="7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.549447 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61"} err="failed to get container status \"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\": rpc error: code = NotFound desc = could not find container \"7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61\": container with ID starting with 7410b8517350d23ee7f979945c49d1018069907202bf2916e72115d2da777f61 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.549465 4818 scope.go:117] "RemoveContainer" containerID="03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.549876 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0"} err="failed to get container status \"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\": rpc error: code = NotFound desc = could not find container \"03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0\": container with ID starting with 03ffed75b6b6f8c7ce6ce9194c0b2e0f39650f95aeb17b4c47d5129c05b4b6e0 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.549892 4818 scope.go:117] "RemoveContainer" containerID="16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.550275 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16"} err="failed to get container status \"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\": rpc error: code = NotFound desc = could not find container \"16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16\": container with ID starting with 16120b24e6d1ce6653561b0d271284a5a883bf6c32345b6ec8d67aba2ef97a16 not found: ID does not exist" Dec 03 06:39:22 crc kubenswrapper[4818]: I1203 06:39:22.746448 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1e30fe6-0b94-4c95-b941-c0623729e123" path="/var/lib/kubelet/pods/a1e30fe6-0b94-4c95-b941-c0623729e123/volumes" Dec 03 06:39:23 crc kubenswrapper[4818]: I1203 06:39:23.169424 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ctpzs_7b25b836-b639-4111-bcea-af7cc791ea32/kube-multus/2.log" Dec 03 06:39:23 crc kubenswrapper[4818]: I1203 06:39:23.171409 4818 generic.go:334] "Generic (PLEG): container finished" podID="d550a052-662c-48b1-9ef2-087a18ad2b4d" containerID="380706f4aecf711c60ae0f13a617416fa1eec9180abf81b305ea8d1192376bed" exitCode=0 Dec 03 06:39:23 crc kubenswrapper[4818]: I1203 06:39:23.171475 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" event={"ID":"d550a052-662c-48b1-9ef2-087a18ad2b4d","Type":"ContainerDied","Data":"380706f4aecf711c60ae0f13a617416fa1eec9180abf81b305ea8d1192376bed"} Dec 03 06:39:23 crc kubenswrapper[4818]: I1203 06:39:23.171519 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" event={"ID":"d550a052-662c-48b1-9ef2-087a18ad2b4d","Type":"ContainerStarted","Data":"c83997a0086c6ba33d448da0ca3286e9b50bc79460c35c1b7876458e6984414a"} Dec 03 06:39:24 crc kubenswrapper[4818]: I1203 06:39:24.182041 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" event={"ID":"d550a052-662c-48b1-9ef2-087a18ad2b4d","Type":"ContainerStarted","Data":"cc9e9b7fa64a2e8324178d10a33cd2c99d64b0385bfd2134e4361396dac8bd0b"} Dec 03 06:39:24 crc kubenswrapper[4818]: I1203 06:39:24.182435 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" event={"ID":"d550a052-662c-48b1-9ef2-087a18ad2b4d","Type":"ContainerStarted","Data":"89575f9da47bccf3d4642b61ceade5fc9b6e6fbf1addb81367e17443469b3ad5"} Dec 03 06:39:24 crc kubenswrapper[4818]: I1203 06:39:24.182446 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" event={"ID":"d550a052-662c-48b1-9ef2-087a18ad2b4d","Type":"ContainerStarted","Data":"511f71aff6dcccf0283586f4d007ee875174c62c97a83b0816a7d46af2ffcc53"} Dec 03 06:39:24 crc kubenswrapper[4818]: I1203 06:39:24.182455 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" event={"ID":"d550a052-662c-48b1-9ef2-087a18ad2b4d","Type":"ContainerStarted","Data":"aa79350a5bdad4831ba12efeadfd199783f3c44b6890287ad053b75b707315c6"} Dec 03 06:39:24 crc kubenswrapper[4818]: I1203 06:39:24.182463 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" event={"ID":"d550a052-662c-48b1-9ef2-087a18ad2b4d","Type":"ContainerStarted","Data":"5df48326cbcf4797c63c7f999dc41ddf8c5dd9e9756352f9ceb654bab2b02f2b"} Dec 03 06:39:24 crc kubenswrapper[4818]: I1203 06:39:24.182471 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" event={"ID":"d550a052-662c-48b1-9ef2-087a18ad2b4d","Type":"ContainerStarted","Data":"e783c4a71eae6e53da72b433fd5c88c4441946d1c50a6efb0d51a7c6ee7e534a"} Dec 03 06:39:27 crc kubenswrapper[4818]: I1203 06:39:27.215303 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" event={"ID":"d550a052-662c-48b1-9ef2-087a18ad2b4d","Type":"ContainerStarted","Data":"76475ea72357bbbf45ebc8065a38ad788c35f400dfb9163236ce0e06e90ff1f6"} Dec 03 06:39:29 crc kubenswrapper[4818]: I1203 06:39:29.232497 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" event={"ID":"d550a052-662c-48b1-9ef2-087a18ad2b4d","Type":"ContainerStarted","Data":"afdc2dc1e734ec1b9f88233069133abd820075bbc014737d1b224fbf36f013db"} Dec 03 06:39:29 crc kubenswrapper[4818]: I1203 06:39:29.233074 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:29 crc kubenswrapper[4818]: I1203 06:39:29.233138 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:29 crc kubenswrapper[4818]: I1203 06:39:29.233154 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:29 crc kubenswrapper[4818]: I1203 06:39:29.267309 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" podStartSLOduration=8.26729173 podStartE2EDuration="8.26729173s" podCreationTimestamp="2025-12-03 06:39:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:39:29.264548291 +0000 UTC m=+726.956157043" watchObservedRunningTime="2025-12-03 06:39:29.26729173 +0000 UTC m=+726.958900482" Dec 03 06:39:29 crc kubenswrapper[4818]: I1203 06:39:29.271377 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:29 crc kubenswrapper[4818]: I1203 06:39:29.272678 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:34 crc kubenswrapper[4818]: I1203 06:39:34.738548 4818 scope.go:117] "RemoveContainer" containerID="3860840e364d23f4d54acf2e87b4f131ee6d2bbd80175613d39d96953478088c" Dec 03 06:39:35 crc kubenswrapper[4818]: I1203 06:39:35.273225 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ctpzs_7b25b836-b639-4111-bcea-af7cc791ea32/kube-multus/2.log" Dec 03 06:39:35 crc kubenswrapper[4818]: I1203 06:39:35.273663 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ctpzs" event={"ID":"7b25b836-b639-4111-bcea-af7cc791ea32","Type":"ContainerStarted","Data":"547d44bd7b6d61aed5deffb74bc6cf567de468c36de59857682f502e2d0b1edf"} Dec 03 06:39:43 crc kubenswrapper[4818]: I1203 06:39:43.302279 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:39:43 crc kubenswrapper[4818]: I1203 06:39:43.303032 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:39:45 crc kubenswrapper[4818]: I1203 06:39:45.761262 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5"] Dec 03 06:39:45 crc kubenswrapper[4818]: I1203 06:39:45.763213 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" Dec 03 06:39:45 crc kubenswrapper[4818]: I1203 06:39:45.767617 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5"] Dec 03 06:39:45 crc kubenswrapper[4818]: I1203 06:39:45.770656 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 06:39:45 crc kubenswrapper[4818]: I1203 06:39:45.857263 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trtjc\" (UniqueName: \"kubernetes.io/projected/37089b98-d9bb-4265-9e86-eeb0e4e0be99-kube-api-access-trtjc\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5\" (UID: \"37089b98-d9bb-4265-9e86-eeb0e4e0be99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" Dec 03 06:39:45 crc kubenswrapper[4818]: I1203 06:39:45.857340 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/37089b98-d9bb-4265-9e86-eeb0e4e0be99-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5\" (UID: \"37089b98-d9bb-4265-9e86-eeb0e4e0be99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" Dec 03 06:39:45 crc kubenswrapper[4818]: I1203 06:39:45.857507 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/37089b98-d9bb-4265-9e86-eeb0e4e0be99-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5\" (UID: \"37089b98-d9bb-4265-9e86-eeb0e4e0be99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" Dec 03 06:39:45 crc kubenswrapper[4818]: I1203 06:39:45.959365 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/37089b98-d9bb-4265-9e86-eeb0e4e0be99-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5\" (UID: \"37089b98-d9bb-4265-9e86-eeb0e4e0be99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" Dec 03 06:39:45 crc kubenswrapper[4818]: I1203 06:39:45.959848 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trtjc\" (UniqueName: \"kubernetes.io/projected/37089b98-d9bb-4265-9e86-eeb0e4e0be99-kube-api-access-trtjc\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5\" (UID: \"37089b98-d9bb-4265-9e86-eeb0e4e0be99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" Dec 03 06:39:45 crc kubenswrapper[4818]: I1203 06:39:45.959911 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/37089b98-d9bb-4265-9e86-eeb0e4e0be99-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5\" (UID: \"37089b98-d9bb-4265-9e86-eeb0e4e0be99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" Dec 03 06:39:45 crc kubenswrapper[4818]: I1203 06:39:45.960214 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/37089b98-d9bb-4265-9e86-eeb0e4e0be99-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5\" (UID: \"37089b98-d9bb-4265-9e86-eeb0e4e0be99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" Dec 03 06:39:45 crc kubenswrapper[4818]: I1203 06:39:45.960636 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/37089b98-d9bb-4265-9e86-eeb0e4e0be99-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5\" (UID: \"37089b98-d9bb-4265-9e86-eeb0e4e0be99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" Dec 03 06:39:45 crc kubenswrapper[4818]: I1203 06:39:45.984431 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trtjc\" (UniqueName: \"kubernetes.io/projected/37089b98-d9bb-4265-9e86-eeb0e4e0be99-kube-api-access-trtjc\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5\" (UID: \"37089b98-d9bb-4265-9e86-eeb0e4e0be99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" Dec 03 06:39:46 crc kubenswrapper[4818]: I1203 06:39:46.083937 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" Dec 03 06:39:46 crc kubenswrapper[4818]: I1203 06:39:46.587197 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5"] Dec 03 06:39:46 crc kubenswrapper[4818]: W1203 06:39:46.594839 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37089b98_d9bb_4265_9e86_eeb0e4e0be99.slice/crio-5cae7c175446cdf7a5dfcb98b9047da8f57c519efcc8201470d47a7f19c21760 WatchSource:0}: Error finding container 5cae7c175446cdf7a5dfcb98b9047da8f57c519efcc8201470d47a7f19c21760: Status 404 returned error can't find the container with id 5cae7c175446cdf7a5dfcb98b9047da8f57c519efcc8201470d47a7f19c21760 Dec 03 06:39:47 crc kubenswrapper[4818]: I1203 06:39:47.360637 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" event={"ID":"37089b98-d9bb-4265-9e86-eeb0e4e0be99","Type":"ContainerStarted","Data":"4449e83b0f3a221d8c648a96cdcff1436ef96517669b19e865dabb4a6cd81d5e"} Dec 03 06:39:47 crc kubenswrapper[4818]: I1203 06:39:47.361097 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" event={"ID":"37089b98-d9bb-4265-9e86-eeb0e4e0be99","Type":"ContainerStarted","Data":"5cae7c175446cdf7a5dfcb98b9047da8f57c519efcc8201470d47a7f19c21760"} Dec 03 06:39:48 crc kubenswrapper[4818]: I1203 06:39:48.366978 4818 generic.go:334] "Generic (PLEG): container finished" podID="37089b98-d9bb-4265-9e86-eeb0e4e0be99" containerID="4449e83b0f3a221d8c648a96cdcff1436ef96517669b19e865dabb4a6cd81d5e" exitCode=0 Dec 03 06:39:48 crc kubenswrapper[4818]: I1203 06:39:48.367036 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" event={"ID":"37089b98-d9bb-4265-9e86-eeb0e4e0be99","Type":"ContainerDied","Data":"4449e83b0f3a221d8c648a96cdcff1436ef96517669b19e865dabb4a6cd81d5e"} Dec 03 06:39:50 crc kubenswrapper[4818]: I1203 06:39:50.384135 4818 generic.go:334] "Generic (PLEG): container finished" podID="37089b98-d9bb-4265-9e86-eeb0e4e0be99" containerID="d9e16eca585f028cb20c23528e5cc0fc8f8c5843eef51f28d629eed95e83eade" exitCode=0 Dec 03 06:39:50 crc kubenswrapper[4818]: I1203 06:39:50.384585 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" event={"ID":"37089b98-d9bb-4265-9e86-eeb0e4e0be99","Type":"ContainerDied","Data":"d9e16eca585f028cb20c23528e5cc0fc8f8c5843eef51f28d629eed95e83eade"} Dec 03 06:39:51 crc kubenswrapper[4818]: I1203 06:39:51.395390 4818 generic.go:334] "Generic (PLEG): container finished" podID="37089b98-d9bb-4265-9e86-eeb0e4e0be99" containerID="b4f491c30a975af82bbed11683e79dff6373e2e2b59cc357985ae59cfed45973" exitCode=0 Dec 03 06:39:51 crc kubenswrapper[4818]: I1203 06:39:51.395481 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" event={"ID":"37089b98-d9bb-4265-9e86-eeb0e4e0be99","Type":"ContainerDied","Data":"b4f491c30a975af82bbed11683e79dff6373e2e2b59cc357985ae59cfed45973"} Dec 03 06:39:52 crc kubenswrapper[4818]: I1203 06:39:52.268143 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gzr42" Dec 03 06:39:52 crc kubenswrapper[4818]: I1203 06:39:52.635575 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" Dec 03 06:39:52 crc kubenswrapper[4818]: I1203 06:39:52.745948 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/37089b98-d9bb-4265-9e86-eeb0e4e0be99-bundle\") pod \"37089b98-d9bb-4265-9e86-eeb0e4e0be99\" (UID: \"37089b98-d9bb-4265-9e86-eeb0e4e0be99\") " Dec 03 06:39:52 crc kubenswrapper[4818]: I1203 06:39:52.746002 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/37089b98-d9bb-4265-9e86-eeb0e4e0be99-util\") pod \"37089b98-d9bb-4265-9e86-eeb0e4e0be99\" (UID: \"37089b98-d9bb-4265-9e86-eeb0e4e0be99\") " Dec 03 06:39:52 crc kubenswrapper[4818]: I1203 06:39:52.746035 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trtjc\" (UniqueName: \"kubernetes.io/projected/37089b98-d9bb-4265-9e86-eeb0e4e0be99-kube-api-access-trtjc\") pod \"37089b98-d9bb-4265-9e86-eeb0e4e0be99\" (UID: \"37089b98-d9bb-4265-9e86-eeb0e4e0be99\") " Dec 03 06:39:52 crc kubenswrapper[4818]: I1203 06:39:52.746971 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37089b98-d9bb-4265-9e86-eeb0e4e0be99-bundle" (OuterVolumeSpecName: "bundle") pod "37089b98-d9bb-4265-9e86-eeb0e4e0be99" (UID: "37089b98-d9bb-4265-9e86-eeb0e4e0be99"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:39:52 crc kubenswrapper[4818]: I1203 06:39:52.751946 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37089b98-d9bb-4265-9e86-eeb0e4e0be99-kube-api-access-trtjc" (OuterVolumeSpecName: "kube-api-access-trtjc") pod "37089b98-d9bb-4265-9e86-eeb0e4e0be99" (UID: "37089b98-d9bb-4265-9e86-eeb0e4e0be99"). InnerVolumeSpecName "kube-api-access-trtjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:39:52 crc kubenswrapper[4818]: I1203 06:39:52.760430 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37089b98-d9bb-4265-9e86-eeb0e4e0be99-util" (OuterVolumeSpecName: "util") pod "37089b98-d9bb-4265-9e86-eeb0e4e0be99" (UID: "37089b98-d9bb-4265-9e86-eeb0e4e0be99"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:39:52 crc kubenswrapper[4818]: I1203 06:39:52.848192 4818 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/37089b98-d9bb-4265-9e86-eeb0e4e0be99-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:52 crc kubenswrapper[4818]: I1203 06:39:52.848238 4818 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/37089b98-d9bb-4265-9e86-eeb0e4e0be99-util\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:52 crc kubenswrapper[4818]: I1203 06:39:52.848255 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trtjc\" (UniqueName: \"kubernetes.io/projected/37089b98-d9bb-4265-9e86-eeb0e4e0be99-kube-api-access-trtjc\") on node \"crc\" DevicePath \"\"" Dec 03 06:39:53 crc kubenswrapper[4818]: I1203 06:39:53.414926 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" event={"ID":"37089b98-d9bb-4265-9e86-eeb0e4e0be99","Type":"ContainerDied","Data":"5cae7c175446cdf7a5dfcb98b9047da8f57c519efcc8201470d47a7f19c21760"} Dec 03 06:39:53 crc kubenswrapper[4818]: I1203 06:39:53.414978 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cae7c175446cdf7a5dfcb98b9047da8f57c519efcc8201470d47a7f19c21760" Dec 03 06:39:53 crc kubenswrapper[4818]: I1203 06:39:53.415027 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5" Dec 03 06:39:54 crc kubenswrapper[4818]: I1203 06:39:54.617866 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-phlrm"] Dec 03 06:39:54 crc kubenswrapper[4818]: E1203 06:39:54.618357 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37089b98-d9bb-4265-9e86-eeb0e4e0be99" containerName="pull" Dec 03 06:39:54 crc kubenswrapper[4818]: I1203 06:39:54.618371 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="37089b98-d9bb-4265-9e86-eeb0e4e0be99" containerName="pull" Dec 03 06:39:54 crc kubenswrapper[4818]: E1203 06:39:54.618391 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37089b98-d9bb-4265-9e86-eeb0e4e0be99" containerName="util" Dec 03 06:39:54 crc kubenswrapper[4818]: I1203 06:39:54.618398 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="37089b98-d9bb-4265-9e86-eeb0e4e0be99" containerName="util" Dec 03 06:39:54 crc kubenswrapper[4818]: E1203 06:39:54.618406 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37089b98-d9bb-4265-9e86-eeb0e4e0be99" containerName="extract" Dec 03 06:39:54 crc kubenswrapper[4818]: I1203 06:39:54.618412 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="37089b98-d9bb-4265-9e86-eeb0e4e0be99" containerName="extract" Dec 03 06:39:54 crc kubenswrapper[4818]: I1203 06:39:54.618505 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="37089b98-d9bb-4265-9e86-eeb0e4e0be99" containerName="extract" Dec 03 06:39:54 crc kubenswrapper[4818]: I1203 06:39:54.618927 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-phlrm" Dec 03 06:39:54 crc kubenswrapper[4818]: I1203 06:39:54.622677 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 03 06:39:54 crc kubenswrapper[4818]: I1203 06:39:54.623418 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 03 06:39:54 crc kubenswrapper[4818]: I1203 06:39:54.633083 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-xqv9t" Dec 03 06:39:54 crc kubenswrapper[4818]: I1203 06:39:54.672447 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcktq\" (UniqueName: \"kubernetes.io/projected/160775bb-417d-4852-bc1f-77d6f8542fad-kube-api-access-bcktq\") pod \"nmstate-operator-5b5b58f5c8-phlrm\" (UID: \"160775bb-417d-4852-bc1f-77d6f8542fad\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-phlrm" Dec 03 06:39:54 crc kubenswrapper[4818]: I1203 06:39:54.690231 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-phlrm"] Dec 03 06:39:54 crc kubenswrapper[4818]: I1203 06:39:54.774057 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcktq\" (UniqueName: \"kubernetes.io/projected/160775bb-417d-4852-bc1f-77d6f8542fad-kube-api-access-bcktq\") pod \"nmstate-operator-5b5b58f5c8-phlrm\" (UID: \"160775bb-417d-4852-bc1f-77d6f8542fad\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-phlrm" Dec 03 06:39:54 crc kubenswrapper[4818]: I1203 06:39:54.788894 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcktq\" (UniqueName: \"kubernetes.io/projected/160775bb-417d-4852-bc1f-77d6f8542fad-kube-api-access-bcktq\") pod \"nmstate-operator-5b5b58f5c8-phlrm\" (UID: \"160775bb-417d-4852-bc1f-77d6f8542fad\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-phlrm" Dec 03 06:39:54 crc kubenswrapper[4818]: I1203 06:39:54.936060 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-phlrm" Dec 03 06:39:55 crc kubenswrapper[4818]: I1203 06:39:55.123915 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-phlrm"] Dec 03 06:39:55 crc kubenswrapper[4818]: W1203 06:39:55.130322 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod160775bb_417d_4852_bc1f_77d6f8542fad.slice/crio-82bd40e856a1000bf77b798801de6bd66e71460628bc5a9829d6629b6cefcbd3 WatchSource:0}: Error finding container 82bd40e856a1000bf77b798801de6bd66e71460628bc5a9829d6629b6cefcbd3: Status 404 returned error can't find the container with id 82bd40e856a1000bf77b798801de6bd66e71460628bc5a9829d6629b6cefcbd3 Dec 03 06:39:55 crc kubenswrapper[4818]: I1203 06:39:55.427521 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-phlrm" event={"ID":"160775bb-417d-4852-bc1f-77d6f8542fad","Type":"ContainerStarted","Data":"82bd40e856a1000bf77b798801de6bd66e71460628bc5a9829d6629b6cefcbd3"} Dec 03 06:39:57 crc kubenswrapper[4818]: I1203 06:39:57.441372 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-phlrm" event={"ID":"160775bb-417d-4852-bc1f-77d6f8542fad","Type":"ContainerStarted","Data":"83e2aed61f717027866b2e1b06b54abb8d2f1e5f148c8e5fc5492ca356643a4b"} Dec 03 06:39:57 crc kubenswrapper[4818]: I1203 06:39:57.472787 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-phlrm" podStartSLOduration=1.3313639 podStartE2EDuration="3.472768024s" podCreationTimestamp="2025-12-03 06:39:54 +0000 UTC" firstStartedPulling="2025-12-03 06:39:55.132178728 +0000 UTC m=+752.823787480" lastFinishedPulling="2025-12-03 06:39:57.273582852 +0000 UTC m=+754.965191604" observedRunningTime="2025-12-03 06:39:57.469217956 +0000 UTC m=+755.160826708" watchObservedRunningTime="2025-12-03 06:39:57.472768024 +0000 UTC m=+755.164376776" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.426393 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-mrrpc"] Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.427538 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mrrpc" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.430369 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-8rfzl" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.433389 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lp854"] Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.434132 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lp854" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.437086 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.447300 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lp854"] Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.464009 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-mrrpc"] Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.504434 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-7hwmb"] Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.505107 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-7hwmb" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.514884 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fbefe4c2-3aea-4974-941e-74fc9448cf32-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-lp854\" (UID: \"fbefe4c2-3aea-4974-941e-74fc9448cf32\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lp854" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.514924 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1d3707ea-c681-45e2-959f-336f11ad216d-nmstate-lock\") pod \"nmstate-handler-7hwmb\" (UID: \"1d3707ea-c681-45e2-959f-336f11ad216d\") " pod="openshift-nmstate/nmstate-handler-7hwmb" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.514944 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k54pv\" (UniqueName: \"kubernetes.io/projected/fbefe4c2-3aea-4974-941e-74fc9448cf32-kube-api-access-k54pv\") pod \"nmstate-webhook-5f6d4c5ccb-lp854\" (UID: \"fbefe4c2-3aea-4974-941e-74fc9448cf32\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lp854" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.514965 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx7sx\" (UniqueName: \"kubernetes.io/projected/1d3707ea-c681-45e2-959f-336f11ad216d-kube-api-access-bx7sx\") pod \"nmstate-handler-7hwmb\" (UID: \"1d3707ea-c681-45e2-959f-336f11ad216d\") " pod="openshift-nmstate/nmstate-handler-7hwmb" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.514989 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1d3707ea-c681-45e2-959f-336f11ad216d-dbus-socket\") pod \"nmstate-handler-7hwmb\" (UID: \"1d3707ea-c681-45e2-959f-336f11ad216d\") " pod="openshift-nmstate/nmstate-handler-7hwmb" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.515021 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1d3707ea-c681-45e2-959f-336f11ad216d-ovs-socket\") pod \"nmstate-handler-7hwmb\" (UID: \"1d3707ea-c681-45e2-959f-336f11ad216d\") " pod="openshift-nmstate/nmstate-handler-7hwmb" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.515053 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv4z2\" (UniqueName: \"kubernetes.io/projected/2ec105b0-3a09-4c8d-a494-95116b4329f2-kube-api-access-sv4z2\") pod \"nmstate-metrics-7f946cbc9-mrrpc\" (UID: \"2ec105b0-3a09-4c8d-a494-95116b4329f2\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mrrpc" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.573803 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-ncmjw"] Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.574578 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-ncmjw" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.576610 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.576678 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-s967f" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.576707 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.589189 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-ncmjw"] Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.615732 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1d3707ea-c681-45e2-959f-336f11ad216d-dbus-socket\") pod \"nmstate-handler-7hwmb\" (UID: \"1d3707ea-c681-45e2-959f-336f11ad216d\") " pod="openshift-nmstate/nmstate-handler-7hwmb" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.615802 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1d3707ea-c681-45e2-959f-336f11ad216d-ovs-socket\") pod \"nmstate-handler-7hwmb\" (UID: \"1d3707ea-c681-45e2-959f-336f11ad216d\") " pod="openshift-nmstate/nmstate-handler-7hwmb" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.615866 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv4z2\" (UniqueName: \"kubernetes.io/projected/2ec105b0-3a09-4c8d-a494-95116b4329f2-kube-api-access-sv4z2\") pod \"nmstate-metrics-7f946cbc9-mrrpc\" (UID: \"2ec105b0-3a09-4c8d-a494-95116b4329f2\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mrrpc" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.615913 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fbefe4c2-3aea-4974-941e-74fc9448cf32-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-lp854\" (UID: \"fbefe4c2-3aea-4974-941e-74fc9448cf32\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lp854" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.615935 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1d3707ea-c681-45e2-959f-336f11ad216d-nmstate-lock\") pod \"nmstate-handler-7hwmb\" (UID: \"1d3707ea-c681-45e2-959f-336f11ad216d\") " pod="openshift-nmstate/nmstate-handler-7hwmb" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.615953 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k54pv\" (UniqueName: \"kubernetes.io/projected/fbefe4c2-3aea-4974-941e-74fc9448cf32-kube-api-access-k54pv\") pod \"nmstate-webhook-5f6d4c5ccb-lp854\" (UID: \"fbefe4c2-3aea-4974-941e-74fc9448cf32\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lp854" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.615975 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx7sx\" (UniqueName: \"kubernetes.io/projected/1d3707ea-c681-45e2-959f-336f11ad216d-kube-api-access-bx7sx\") pod \"nmstate-handler-7hwmb\" (UID: \"1d3707ea-c681-45e2-959f-336f11ad216d\") " pod="openshift-nmstate/nmstate-handler-7hwmb" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.616246 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1d3707ea-c681-45e2-959f-336f11ad216d-dbus-socket\") pod \"nmstate-handler-7hwmb\" (UID: \"1d3707ea-c681-45e2-959f-336f11ad216d\") " pod="openshift-nmstate/nmstate-handler-7hwmb" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.616354 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1d3707ea-c681-45e2-959f-336f11ad216d-nmstate-lock\") pod \"nmstate-handler-7hwmb\" (UID: \"1d3707ea-c681-45e2-959f-336f11ad216d\") " pod="openshift-nmstate/nmstate-handler-7hwmb" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.616518 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1d3707ea-c681-45e2-959f-336f11ad216d-ovs-socket\") pod \"nmstate-handler-7hwmb\" (UID: \"1d3707ea-c681-45e2-959f-336f11ad216d\") " pod="openshift-nmstate/nmstate-handler-7hwmb" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.621541 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/fbefe4c2-3aea-4974-941e-74fc9448cf32-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-lp854\" (UID: \"fbefe4c2-3aea-4974-941e-74fc9448cf32\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lp854" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.631557 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv4z2\" (UniqueName: \"kubernetes.io/projected/2ec105b0-3a09-4c8d-a494-95116b4329f2-kube-api-access-sv4z2\") pod \"nmstate-metrics-7f946cbc9-mrrpc\" (UID: \"2ec105b0-3a09-4c8d-a494-95116b4329f2\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mrrpc" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.632558 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx7sx\" (UniqueName: \"kubernetes.io/projected/1d3707ea-c681-45e2-959f-336f11ad216d-kube-api-access-bx7sx\") pod \"nmstate-handler-7hwmb\" (UID: \"1d3707ea-c681-45e2-959f-336f11ad216d\") " pod="openshift-nmstate/nmstate-handler-7hwmb" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.637452 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k54pv\" (UniqueName: \"kubernetes.io/projected/fbefe4c2-3aea-4974-941e-74fc9448cf32-kube-api-access-k54pv\") pod \"nmstate-webhook-5f6d4c5ccb-lp854\" (UID: \"fbefe4c2-3aea-4974-941e-74fc9448cf32\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lp854" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.717617 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3c885e5f-3396-4927-90ef-878b255a6317-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-ncmjw\" (UID: \"3c885e5f-3396-4927-90ef-878b255a6317\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-ncmjw" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.717672 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3c885e5f-3396-4927-90ef-878b255a6317-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-ncmjw\" (UID: \"3c885e5f-3396-4927-90ef-878b255a6317\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-ncmjw" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.717708 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbsgj\" (UniqueName: \"kubernetes.io/projected/3c885e5f-3396-4927-90ef-878b255a6317-kube-api-access-pbsgj\") pod \"nmstate-console-plugin-7fbb5f6569-ncmjw\" (UID: \"3c885e5f-3396-4927-90ef-878b255a6317\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-ncmjw" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.768020 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-669c77b7c7-q8hjr"] Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.768951 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.780244 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-669c77b7c7-q8hjr"] Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.805137 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mrrpc" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.818191 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lp854" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.818565 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3c885e5f-3396-4927-90ef-878b255a6317-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-ncmjw\" (UID: \"3c885e5f-3396-4927-90ef-878b255a6317\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-ncmjw" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.818615 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbsgj\" (UniqueName: \"kubernetes.io/projected/3c885e5f-3396-4927-90ef-878b255a6317-kube-api-access-pbsgj\") pod \"nmstate-console-plugin-7fbb5f6569-ncmjw\" (UID: \"3c885e5f-3396-4927-90ef-878b255a6317\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-ncmjw" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.818685 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3c885e5f-3396-4927-90ef-878b255a6317-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-ncmjw\" (UID: \"3c885e5f-3396-4927-90ef-878b255a6317\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-ncmjw" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.819572 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3c885e5f-3396-4927-90ef-878b255a6317-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-ncmjw\" (UID: \"3c885e5f-3396-4927-90ef-878b255a6317\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-ncmjw" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.823382 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3c885e5f-3396-4927-90ef-878b255a6317-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-ncmjw\" (UID: \"3c885e5f-3396-4927-90ef-878b255a6317\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-ncmjw" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.830263 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-7hwmb" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.835176 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbsgj\" (UniqueName: \"kubernetes.io/projected/3c885e5f-3396-4927-90ef-878b255a6317-kube-api-access-pbsgj\") pod \"nmstate-console-plugin-7fbb5f6569-ncmjw\" (UID: \"3c885e5f-3396-4927-90ef-878b255a6317\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-ncmjw" Dec 03 06:39:58 crc kubenswrapper[4818]: W1203 06:39:58.870677 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d3707ea_c681_45e2_959f_336f11ad216d.slice/crio-575945da48a09aedf31552d4d75ddecc1b6780fda01c0228dd5a9ffe772bcc1c WatchSource:0}: Error finding container 575945da48a09aedf31552d4d75ddecc1b6780fda01c0228dd5a9ffe772bcc1c: Status 404 returned error can't find the container with id 575945da48a09aedf31552d4d75ddecc1b6780fda01c0228dd5a9ffe772bcc1c Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.910019 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-ncmjw" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.919273 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/52daf3cd-54f4-463b-bd3c-78e5a5341858-console-oauth-config\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.919330 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/52daf3cd-54f4-463b-bd3c-78e5a5341858-oauth-serving-cert\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.919351 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/52daf3cd-54f4-463b-bd3c-78e5a5341858-console-serving-cert\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.919571 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/52daf3cd-54f4-463b-bd3c-78e5a5341858-console-config\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.919685 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7qf9\" (UniqueName: \"kubernetes.io/projected/52daf3cd-54f4-463b-bd3c-78e5a5341858-kube-api-access-z7qf9\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.921188 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52daf3cd-54f4-463b-bd3c-78e5a5341858-trusted-ca-bundle\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:58 crc kubenswrapper[4818]: I1203 06:39:58.921222 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/52daf3cd-54f4-463b-bd3c-78e5a5341858-service-ca\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.022108 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/52daf3cd-54f4-463b-bd3c-78e5a5341858-console-config\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.022157 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7qf9\" (UniqueName: \"kubernetes.io/projected/52daf3cd-54f4-463b-bd3c-78e5a5341858-kube-api-access-z7qf9\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.022204 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52daf3cd-54f4-463b-bd3c-78e5a5341858-trusted-ca-bundle\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.022230 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/52daf3cd-54f4-463b-bd3c-78e5a5341858-service-ca\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.022279 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/52daf3cd-54f4-463b-bd3c-78e5a5341858-console-oauth-config\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.022312 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/52daf3cd-54f4-463b-bd3c-78e5a5341858-oauth-serving-cert\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.022331 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/52daf3cd-54f4-463b-bd3c-78e5a5341858-console-serving-cert\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.023490 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/52daf3cd-54f4-463b-bd3c-78e5a5341858-service-ca\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.023617 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52daf3cd-54f4-463b-bd3c-78e5a5341858-trusted-ca-bundle\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.024092 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/52daf3cd-54f4-463b-bd3c-78e5a5341858-oauth-serving-cert\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.024506 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/52daf3cd-54f4-463b-bd3c-78e5a5341858-console-config\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.028504 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/52daf3cd-54f4-463b-bd3c-78e5a5341858-console-serving-cert\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.037401 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/52daf3cd-54f4-463b-bd3c-78e5a5341858-console-oauth-config\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.039667 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7qf9\" (UniqueName: \"kubernetes.io/projected/52daf3cd-54f4-463b-bd3c-78e5a5341858-kube-api-access-z7qf9\") pod \"console-669c77b7c7-q8hjr\" (UID: \"52daf3cd-54f4-463b-bd3c-78e5a5341858\") " pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.090326 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.094154 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lp854"] Dec 03 06:39:59 crc kubenswrapper[4818]: W1203 06:39:59.100527 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbefe4c2_3aea_4974_941e_74fc9448cf32.slice/crio-75d9cf405d063c1d96a68036e78be3608ffce50361ba2c01f51e5fdf62f7f2d6 WatchSource:0}: Error finding container 75d9cf405d063c1d96a68036e78be3608ffce50361ba2c01f51e5fdf62f7f2d6: Status 404 returned error can't find the container with id 75d9cf405d063c1d96a68036e78be3608ffce50361ba2c01f51e5fdf62f7f2d6 Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.161034 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-ncmjw"] Dec 03 06:39:59 crc kubenswrapper[4818]: W1203 06:39:59.165224 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c885e5f_3396_4927_90ef_878b255a6317.slice/crio-b2f8849c9101bde27e265f54ac89c5e98ef7a9dfcd554816289c523ab46e9761 WatchSource:0}: Error finding container b2f8849c9101bde27e265f54ac89c5e98ef7a9dfcd554816289c523ab46e9761: Status 404 returned error can't find the container with id b2f8849c9101bde27e265f54ac89c5e98ef7a9dfcd554816289c523ab46e9761 Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.245691 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-mrrpc"] Dec 03 06:39:59 crc kubenswrapper[4818]: W1203 06:39:59.252679 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ec105b0_3a09_4c8d_a494_95116b4329f2.slice/crio-2e128bbd8b1635a397345c72b5e9a7aaecd2e20baa2a8d5db8518986aa90f5fb WatchSource:0}: Error finding container 2e128bbd8b1635a397345c72b5e9a7aaecd2e20baa2a8d5db8518986aa90f5fb: Status 404 returned error can't find the container with id 2e128bbd8b1635a397345c72b5e9a7aaecd2e20baa2a8d5db8518986aa90f5fb Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.306409 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-669c77b7c7-q8hjr"] Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.455726 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-ncmjw" event={"ID":"3c885e5f-3396-4927-90ef-878b255a6317","Type":"ContainerStarted","Data":"b2f8849c9101bde27e265f54ac89c5e98ef7a9dfcd554816289c523ab46e9761"} Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.457950 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-669c77b7c7-q8hjr" event={"ID":"52daf3cd-54f4-463b-bd3c-78e5a5341858","Type":"ContainerStarted","Data":"f8f82bebc89b9116db7c8b69125d6db97d8655c981302db63558328cb541a4f3"} Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.457995 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-669c77b7c7-q8hjr" event={"ID":"52daf3cd-54f4-463b-bd3c-78e5a5341858","Type":"ContainerStarted","Data":"82aa8a9b6aaca1807dbb9a970fd74eba1d918cdbe3dbcecb605b75e6e751de7e"} Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.460363 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-7hwmb" event={"ID":"1d3707ea-c681-45e2-959f-336f11ad216d","Type":"ContainerStarted","Data":"575945da48a09aedf31552d4d75ddecc1b6780fda01c0228dd5a9ffe772bcc1c"} Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.463578 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lp854" event={"ID":"fbefe4c2-3aea-4974-941e-74fc9448cf32","Type":"ContainerStarted","Data":"75d9cf405d063c1d96a68036e78be3608ffce50361ba2c01f51e5fdf62f7f2d6"} Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.464593 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mrrpc" event={"ID":"2ec105b0-3a09-4c8d-a494-95116b4329f2","Type":"ContainerStarted","Data":"2e128bbd8b1635a397345c72b5e9a7aaecd2e20baa2a8d5db8518986aa90f5fb"} Dec 03 06:39:59 crc kubenswrapper[4818]: I1203 06:39:59.484335 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-669c77b7c7-q8hjr" podStartSLOduration=1.484319165 podStartE2EDuration="1.484319165s" podCreationTimestamp="2025-12-03 06:39:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:39:59.483481304 +0000 UTC m=+757.175090056" watchObservedRunningTime="2025-12-03 06:39:59.484319165 +0000 UTC m=+757.175927917" Dec 03 06:40:02 crc kubenswrapper[4818]: I1203 06:40:02.485185 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-7hwmb" event={"ID":"1d3707ea-c681-45e2-959f-336f11ad216d","Type":"ContainerStarted","Data":"1509a93318e86be383868b3eef0ac71de084eb2281052de78b54c627c402a45a"} Dec 03 06:40:02 crc kubenswrapper[4818]: I1203 06:40:02.485793 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-7hwmb" Dec 03 06:40:02 crc kubenswrapper[4818]: I1203 06:40:02.486633 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lp854" event={"ID":"fbefe4c2-3aea-4974-941e-74fc9448cf32","Type":"ContainerStarted","Data":"047959c27d6880265fe69a401a52eb91c88c21cf9b36b80f4727880b4f99f3a8"} Dec 03 06:40:02 crc kubenswrapper[4818]: I1203 06:40:02.486717 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lp854" Dec 03 06:40:02 crc kubenswrapper[4818]: I1203 06:40:02.488021 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mrrpc" event={"ID":"2ec105b0-3a09-4c8d-a494-95116b4329f2","Type":"ContainerStarted","Data":"a337aeb78ebf966cb7bb730a864148fdf5fc2a9e4001dd15635a4c6c6e34c788"} Dec 03 06:40:02 crc kubenswrapper[4818]: I1203 06:40:02.489658 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-ncmjw" event={"ID":"3c885e5f-3396-4927-90ef-878b255a6317","Type":"ContainerStarted","Data":"cf9846e83cad397e89a0eb2edd295a07963a9e6d5e0de683093fc0ed4d6cac76"} Dec 03 06:40:02 crc kubenswrapper[4818]: I1203 06:40:02.512468 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-7hwmb" podStartSLOduration=1.320407528 podStartE2EDuration="4.51243972s" podCreationTimestamp="2025-12-03 06:39:58 +0000 UTC" firstStartedPulling="2025-12-03 06:39:58.872410682 +0000 UTC m=+756.564019434" lastFinishedPulling="2025-12-03 06:40:02.064442854 +0000 UTC m=+759.756051626" observedRunningTime="2025-12-03 06:40:02.507295572 +0000 UTC m=+760.198904344" watchObservedRunningTime="2025-12-03 06:40:02.51243972 +0000 UTC m=+760.204048482" Dec 03 06:40:02 crc kubenswrapper[4818]: I1203 06:40:02.537433 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-ncmjw" podStartSLOduration=1.679863338 podStartE2EDuration="4.537401609s" podCreationTimestamp="2025-12-03 06:39:58 +0000 UTC" firstStartedPulling="2025-12-03 06:39:59.175652157 +0000 UTC m=+756.867260909" lastFinishedPulling="2025-12-03 06:40:02.033190398 +0000 UTC m=+759.724799180" observedRunningTime="2025-12-03 06:40:02.527908264 +0000 UTC m=+760.219517016" watchObservedRunningTime="2025-12-03 06:40:02.537401609 +0000 UTC m=+760.229010381" Dec 03 06:40:02 crc kubenswrapper[4818]: I1203 06:40:02.557223 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lp854" podStartSLOduration=1.6270197560000002 podStartE2EDuration="4.557203481s" podCreationTimestamp="2025-12-03 06:39:58 +0000 UTC" firstStartedPulling="2025-12-03 06:39:59.102694106 +0000 UTC m=+756.794302858" lastFinishedPulling="2025-12-03 06:40:02.032877791 +0000 UTC m=+759.724486583" observedRunningTime="2025-12-03 06:40:02.554727109 +0000 UTC m=+760.246335871" watchObservedRunningTime="2025-12-03 06:40:02.557203481 +0000 UTC m=+760.248812243" Dec 03 06:40:05 crc kubenswrapper[4818]: I1203 06:40:05.505305 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mrrpc" event={"ID":"2ec105b0-3a09-4c8d-a494-95116b4329f2","Type":"ContainerStarted","Data":"9b50997a1e945f98b0cfb7bae23228aeb81f154890cf305900dcea0724973663"} Dec 03 06:40:05 crc kubenswrapper[4818]: I1203 06:40:05.518486 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mrrpc" podStartSLOduration=1.6000946379999998 podStartE2EDuration="7.518463276s" podCreationTimestamp="2025-12-03 06:39:58 +0000 UTC" firstStartedPulling="2025-12-03 06:39:59.255018706 +0000 UTC m=+756.946627458" lastFinishedPulling="2025-12-03 06:40:05.173387344 +0000 UTC m=+762.864996096" observedRunningTime="2025-12-03 06:40:05.518308822 +0000 UTC m=+763.209917574" watchObservedRunningTime="2025-12-03 06:40:05.518463276 +0000 UTC m=+763.210072028" Dec 03 06:40:08 crc kubenswrapper[4818]: I1203 06:40:08.871351 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-7hwmb" Dec 03 06:40:09 crc kubenswrapper[4818]: I1203 06:40:09.091520 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:40:09 crc kubenswrapper[4818]: I1203 06:40:09.091604 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:40:09 crc kubenswrapper[4818]: I1203 06:40:09.102139 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:40:09 crc kubenswrapper[4818]: I1203 06:40:09.540904 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-669c77b7c7-q8hjr" Dec 03 06:40:09 crc kubenswrapper[4818]: I1203 06:40:09.638508 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-sccrw"] Dec 03 06:40:13 crc kubenswrapper[4818]: I1203 06:40:13.303009 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:40:13 crc kubenswrapper[4818]: I1203 06:40:13.303600 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:40:14 crc kubenswrapper[4818]: I1203 06:40:14.213144 4818 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 06:40:18 crc kubenswrapper[4818]: I1203 06:40:18.827307 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-lp854" Dec 03 06:40:32 crc kubenswrapper[4818]: I1203 06:40:32.869144 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm"] Dec 03 06:40:32 crc kubenswrapper[4818]: I1203 06:40:32.871351 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" Dec 03 06:40:32 crc kubenswrapper[4818]: I1203 06:40:32.880619 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm"] Dec 03 06:40:32 crc kubenswrapper[4818]: I1203 06:40:32.880883 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 06:40:33 crc kubenswrapper[4818]: I1203 06:40:33.011299 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5ln8\" (UniqueName: \"kubernetes.io/projected/9897d8d1-c815-45ee-9c90-0b1e34a71340-kube-api-access-x5ln8\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm\" (UID: \"9897d8d1-c815-45ee-9c90-0b1e34a71340\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" Dec 03 06:40:33 crc kubenswrapper[4818]: I1203 06:40:33.011494 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9897d8d1-c815-45ee-9c90-0b1e34a71340-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm\" (UID: \"9897d8d1-c815-45ee-9c90-0b1e34a71340\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" Dec 03 06:40:33 crc kubenswrapper[4818]: I1203 06:40:33.012038 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9897d8d1-c815-45ee-9c90-0b1e34a71340-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm\" (UID: \"9897d8d1-c815-45ee-9c90-0b1e34a71340\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" Dec 03 06:40:33 crc kubenswrapper[4818]: I1203 06:40:33.113898 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5ln8\" (UniqueName: \"kubernetes.io/projected/9897d8d1-c815-45ee-9c90-0b1e34a71340-kube-api-access-x5ln8\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm\" (UID: \"9897d8d1-c815-45ee-9c90-0b1e34a71340\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" Dec 03 06:40:33 crc kubenswrapper[4818]: I1203 06:40:33.114269 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9897d8d1-c815-45ee-9c90-0b1e34a71340-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm\" (UID: \"9897d8d1-c815-45ee-9c90-0b1e34a71340\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" Dec 03 06:40:33 crc kubenswrapper[4818]: I1203 06:40:33.114354 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9897d8d1-c815-45ee-9c90-0b1e34a71340-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm\" (UID: \"9897d8d1-c815-45ee-9c90-0b1e34a71340\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" Dec 03 06:40:33 crc kubenswrapper[4818]: I1203 06:40:33.114787 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9897d8d1-c815-45ee-9c90-0b1e34a71340-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm\" (UID: \"9897d8d1-c815-45ee-9c90-0b1e34a71340\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" Dec 03 06:40:33 crc kubenswrapper[4818]: I1203 06:40:33.114837 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9897d8d1-c815-45ee-9c90-0b1e34a71340-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm\" (UID: \"9897d8d1-c815-45ee-9c90-0b1e34a71340\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" Dec 03 06:40:33 crc kubenswrapper[4818]: I1203 06:40:33.135290 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5ln8\" (UniqueName: \"kubernetes.io/projected/9897d8d1-c815-45ee-9c90-0b1e34a71340-kube-api-access-x5ln8\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm\" (UID: \"9897d8d1-c815-45ee-9c90-0b1e34a71340\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" Dec 03 06:40:33 crc kubenswrapper[4818]: I1203 06:40:33.200438 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" Dec 03 06:40:33 crc kubenswrapper[4818]: I1203 06:40:33.727154 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm"] Dec 03 06:40:34 crc kubenswrapper[4818]: I1203 06:40:34.690030 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-sccrw" podUID="6d2839d4-e624-4d61-9225-894a998e9e9e" containerName="console" containerID="cri-o://81e079fab9ada721df205aef6940ff13656d56bcf1aebbc1b718d9e56c04f368" gracePeriod=15 Dec 03 06:40:34 crc kubenswrapper[4818]: I1203 06:40:34.719668 4818 generic.go:334] "Generic (PLEG): container finished" podID="9897d8d1-c815-45ee-9c90-0b1e34a71340" containerID="56b03c64d0a663d05ab6b930f996eb04c960b0f55411a54fc7fc1ab8e76390ef" exitCode=0 Dec 03 06:40:34 crc kubenswrapper[4818]: I1203 06:40:34.719727 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" event={"ID":"9897d8d1-c815-45ee-9c90-0b1e34a71340","Type":"ContainerDied","Data":"56b03c64d0a663d05ab6b930f996eb04c960b0f55411a54fc7fc1ab8e76390ef"} Dec 03 06:40:34 crc kubenswrapper[4818]: I1203 06:40:34.719766 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" event={"ID":"9897d8d1-c815-45ee-9c90-0b1e34a71340","Type":"ContainerStarted","Data":"d408b95d7d3c42882853ae9db16d1e70659dc904f7539b3de67851c18f36d6f5"} Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.211141 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-brv8j"] Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.215271 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-brv8j" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.223150 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-brv8j"] Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.223193 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-sccrw_6d2839d4-e624-4d61-9225-894a998e9e9e/console/0.log" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.223302 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.249031 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6d2839d4-e624-4d61-9225-894a998e9e9e-console-serving-cert\") pod \"6d2839d4-e624-4d61-9225-894a998e9e9e\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.249074 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-service-ca\") pod \"6d2839d4-e624-4d61-9225-894a998e9e9e\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.249107 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-trusted-ca-bundle\") pod \"6d2839d4-e624-4d61-9225-894a998e9e9e\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.249139 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6d2839d4-e624-4d61-9225-894a998e9e9e-console-oauth-config\") pod \"6d2839d4-e624-4d61-9225-894a998e9e9e\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.249190 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-oauth-serving-cert\") pod \"6d2839d4-e624-4d61-9225-894a998e9e9e\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.249212 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkks8\" (UniqueName: \"kubernetes.io/projected/6d2839d4-e624-4d61-9225-894a998e9e9e-kube-api-access-vkks8\") pod \"6d2839d4-e624-4d61-9225-894a998e9e9e\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.249259 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-console-config\") pod \"6d2839d4-e624-4d61-9225-894a998e9e9e\" (UID: \"6d2839d4-e624-4d61-9225-894a998e9e9e\") " Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.249357 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df7f10a-718a-45d5-aad1-72fb5871bea8-utilities\") pod \"redhat-operators-brv8j\" (UID: \"1df7f10a-718a-45d5-aad1-72fb5871bea8\") " pod="openshift-marketplace/redhat-operators-brv8j" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.249390 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr6fv\" (UniqueName: \"kubernetes.io/projected/1df7f10a-718a-45d5-aad1-72fb5871bea8-kube-api-access-qr6fv\") pod \"redhat-operators-brv8j\" (UID: \"1df7f10a-718a-45d5-aad1-72fb5871bea8\") " pod="openshift-marketplace/redhat-operators-brv8j" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.249412 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df7f10a-718a-45d5-aad1-72fb5871bea8-catalog-content\") pod \"redhat-operators-brv8j\" (UID: \"1df7f10a-718a-45d5-aad1-72fb5871bea8\") " pod="openshift-marketplace/redhat-operators-brv8j" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.250215 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-console-config" (OuterVolumeSpecName: "console-config") pod "6d2839d4-e624-4d61-9225-894a998e9e9e" (UID: "6d2839d4-e624-4d61-9225-894a998e9e9e"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.250737 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "6d2839d4-e624-4d61-9225-894a998e9e9e" (UID: "6d2839d4-e624-4d61-9225-894a998e9e9e"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.251083 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-service-ca" (OuterVolumeSpecName: "service-ca") pod "6d2839d4-e624-4d61-9225-894a998e9e9e" (UID: "6d2839d4-e624-4d61-9225-894a998e9e9e"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.251083 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6d2839d4-e624-4d61-9225-894a998e9e9e" (UID: "6d2839d4-e624-4d61-9225-894a998e9e9e"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.258350 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d2839d4-e624-4d61-9225-894a998e9e9e-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "6d2839d4-e624-4d61-9225-894a998e9e9e" (UID: "6d2839d4-e624-4d61-9225-894a998e9e9e"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.259262 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d2839d4-e624-4d61-9225-894a998e9e9e-kube-api-access-vkks8" (OuterVolumeSpecName: "kube-api-access-vkks8") pod "6d2839d4-e624-4d61-9225-894a998e9e9e" (UID: "6d2839d4-e624-4d61-9225-894a998e9e9e"). InnerVolumeSpecName "kube-api-access-vkks8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.271833 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d2839d4-e624-4d61-9225-894a998e9e9e-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "6d2839d4-e624-4d61-9225-894a998e9e9e" (UID: "6d2839d4-e624-4d61-9225-894a998e9e9e"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.351045 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr6fv\" (UniqueName: \"kubernetes.io/projected/1df7f10a-718a-45d5-aad1-72fb5871bea8-kube-api-access-qr6fv\") pod \"redhat-operators-brv8j\" (UID: \"1df7f10a-718a-45d5-aad1-72fb5871bea8\") " pod="openshift-marketplace/redhat-operators-brv8j" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.351716 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df7f10a-718a-45d5-aad1-72fb5871bea8-catalog-content\") pod \"redhat-operators-brv8j\" (UID: \"1df7f10a-718a-45d5-aad1-72fb5871bea8\") " pod="openshift-marketplace/redhat-operators-brv8j" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.351779 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df7f10a-718a-45d5-aad1-72fb5871bea8-catalog-content\") pod \"redhat-operators-brv8j\" (UID: \"1df7f10a-718a-45d5-aad1-72fb5871bea8\") " pod="openshift-marketplace/redhat-operators-brv8j" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.351914 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df7f10a-718a-45d5-aad1-72fb5871bea8-utilities\") pod \"redhat-operators-brv8j\" (UID: \"1df7f10a-718a-45d5-aad1-72fb5871bea8\") " pod="openshift-marketplace/redhat-operators-brv8j" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.352008 4818 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.352034 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkks8\" (UniqueName: \"kubernetes.io/projected/6d2839d4-e624-4d61-9225-894a998e9e9e-kube-api-access-vkks8\") on node \"crc\" DevicePath \"\"" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.352046 4818 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.352055 4818 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6d2839d4-e624-4d61-9225-894a998e9e9e-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.352066 4818 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.352077 4818 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6d2839d4-e624-4d61-9225-894a998e9e9e-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.352085 4818 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6d2839d4-e624-4d61-9225-894a998e9e9e-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.352457 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df7f10a-718a-45d5-aad1-72fb5871bea8-utilities\") pod \"redhat-operators-brv8j\" (UID: \"1df7f10a-718a-45d5-aad1-72fb5871bea8\") " pod="openshift-marketplace/redhat-operators-brv8j" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.370498 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr6fv\" (UniqueName: \"kubernetes.io/projected/1df7f10a-718a-45d5-aad1-72fb5871bea8-kube-api-access-qr6fv\") pod \"redhat-operators-brv8j\" (UID: \"1df7f10a-718a-45d5-aad1-72fb5871bea8\") " pod="openshift-marketplace/redhat-operators-brv8j" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.551966 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-brv8j" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.729513 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-sccrw_6d2839d4-e624-4d61-9225-894a998e9e9e/console/0.log" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.730172 4818 generic.go:334] "Generic (PLEG): container finished" podID="6d2839d4-e624-4d61-9225-894a998e9e9e" containerID="81e079fab9ada721df205aef6940ff13656d56bcf1aebbc1b718d9e56c04f368" exitCode=2 Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.730224 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sccrw" event={"ID":"6d2839d4-e624-4d61-9225-894a998e9e9e","Type":"ContainerDied","Data":"81e079fab9ada721df205aef6940ff13656d56bcf1aebbc1b718d9e56c04f368"} Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.730272 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sccrw" event={"ID":"6d2839d4-e624-4d61-9225-894a998e9e9e","Type":"ContainerDied","Data":"2f41b69b4237893c5f9b1a86b6176f06169b181984f8192f8164e15d728b7d65"} Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.730292 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sccrw" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.730299 4818 scope.go:117] "RemoveContainer" containerID="81e079fab9ada721df205aef6940ff13656d56bcf1aebbc1b718d9e56c04f368" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.758316 4818 scope.go:117] "RemoveContainer" containerID="81e079fab9ada721df205aef6940ff13656d56bcf1aebbc1b718d9e56c04f368" Dec 03 06:40:35 crc kubenswrapper[4818]: E1203 06:40:35.758978 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81e079fab9ada721df205aef6940ff13656d56bcf1aebbc1b718d9e56c04f368\": container with ID starting with 81e079fab9ada721df205aef6940ff13656d56bcf1aebbc1b718d9e56c04f368 not found: ID does not exist" containerID="81e079fab9ada721df205aef6940ff13656d56bcf1aebbc1b718d9e56c04f368" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.759010 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81e079fab9ada721df205aef6940ff13656d56bcf1aebbc1b718d9e56c04f368"} err="failed to get container status \"81e079fab9ada721df205aef6940ff13656d56bcf1aebbc1b718d9e56c04f368\": rpc error: code = NotFound desc = could not find container \"81e079fab9ada721df205aef6940ff13656d56bcf1aebbc1b718d9e56c04f368\": container with ID starting with 81e079fab9ada721df205aef6940ff13656d56bcf1aebbc1b718d9e56c04f368 not found: ID does not exist" Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.768938 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-sccrw"] Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.793149 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-sccrw"] Dec 03 06:40:35 crc kubenswrapper[4818]: I1203 06:40:35.875194 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-brv8j"] Dec 03 06:40:36 crc kubenswrapper[4818]: I1203 06:40:36.742206 4818 generic.go:334] "Generic (PLEG): container finished" podID="9897d8d1-c815-45ee-9c90-0b1e34a71340" containerID="63abdc74d14678008b54d9e38b103715482ac1e9f172c64d398c65de4c2b2e02" exitCode=0 Dec 03 06:40:36 crc kubenswrapper[4818]: I1203 06:40:36.744953 4818 generic.go:334] "Generic (PLEG): container finished" podID="1df7f10a-718a-45d5-aad1-72fb5871bea8" containerID="2852a7705fbaa7e5b8d7216abb77d4d6e2995a9ec18394a97fed1b193341a7bd" exitCode=0 Dec 03 06:40:36 crc kubenswrapper[4818]: I1203 06:40:36.749926 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d2839d4-e624-4d61-9225-894a998e9e9e" path="/var/lib/kubelet/pods/6d2839d4-e624-4d61-9225-894a998e9e9e/volumes" Dec 03 06:40:36 crc kubenswrapper[4818]: I1203 06:40:36.750633 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" event={"ID":"9897d8d1-c815-45ee-9c90-0b1e34a71340","Type":"ContainerDied","Data":"63abdc74d14678008b54d9e38b103715482ac1e9f172c64d398c65de4c2b2e02"} Dec 03 06:40:36 crc kubenswrapper[4818]: I1203 06:40:36.750669 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brv8j" event={"ID":"1df7f10a-718a-45d5-aad1-72fb5871bea8","Type":"ContainerDied","Data":"2852a7705fbaa7e5b8d7216abb77d4d6e2995a9ec18394a97fed1b193341a7bd"} Dec 03 06:40:36 crc kubenswrapper[4818]: I1203 06:40:36.750690 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brv8j" event={"ID":"1df7f10a-718a-45d5-aad1-72fb5871bea8","Type":"ContainerStarted","Data":"f604d2396d4833b774661e06fea8a486b13b784511debe336e1d704a9f397033"} Dec 03 06:40:37 crc kubenswrapper[4818]: I1203 06:40:37.761765 4818 generic.go:334] "Generic (PLEG): container finished" podID="9897d8d1-c815-45ee-9c90-0b1e34a71340" containerID="3ddefbfd782a7dfd8534e3fb39721cce5f9e3097ea5f4992b2056abc3671d43f" exitCode=0 Dec 03 06:40:37 crc kubenswrapper[4818]: I1203 06:40:37.761952 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" event={"ID":"9897d8d1-c815-45ee-9c90-0b1e34a71340","Type":"ContainerDied","Data":"3ddefbfd782a7dfd8534e3fb39721cce5f9e3097ea5f4992b2056abc3671d43f"} Dec 03 06:40:37 crc kubenswrapper[4818]: I1203 06:40:37.765001 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brv8j" event={"ID":"1df7f10a-718a-45d5-aad1-72fb5871bea8","Type":"ContainerStarted","Data":"541ec01d3b4a4c50f002dfff30455d5bd209b36f1f116fed6c171ce42abd62a4"} Dec 03 06:40:39 crc kubenswrapper[4818]: I1203 06:40:39.234237 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" Dec 03 06:40:39 crc kubenswrapper[4818]: I1203 06:40:39.408555 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9897d8d1-c815-45ee-9c90-0b1e34a71340-util\") pod \"9897d8d1-c815-45ee-9c90-0b1e34a71340\" (UID: \"9897d8d1-c815-45ee-9c90-0b1e34a71340\") " Dec 03 06:40:39 crc kubenswrapper[4818]: I1203 06:40:39.409442 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9897d8d1-c815-45ee-9c90-0b1e34a71340-bundle\") pod \"9897d8d1-c815-45ee-9c90-0b1e34a71340\" (UID: \"9897d8d1-c815-45ee-9c90-0b1e34a71340\") " Dec 03 06:40:39 crc kubenswrapper[4818]: I1203 06:40:39.409563 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5ln8\" (UniqueName: \"kubernetes.io/projected/9897d8d1-c815-45ee-9c90-0b1e34a71340-kube-api-access-x5ln8\") pod \"9897d8d1-c815-45ee-9c90-0b1e34a71340\" (UID: \"9897d8d1-c815-45ee-9c90-0b1e34a71340\") " Dec 03 06:40:39 crc kubenswrapper[4818]: I1203 06:40:39.410548 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9897d8d1-c815-45ee-9c90-0b1e34a71340-bundle" (OuterVolumeSpecName: "bundle") pod "9897d8d1-c815-45ee-9c90-0b1e34a71340" (UID: "9897d8d1-c815-45ee-9c90-0b1e34a71340"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:40:39 crc kubenswrapper[4818]: I1203 06:40:39.418093 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9897d8d1-c815-45ee-9c90-0b1e34a71340-kube-api-access-x5ln8" (OuterVolumeSpecName: "kube-api-access-x5ln8") pod "9897d8d1-c815-45ee-9c90-0b1e34a71340" (UID: "9897d8d1-c815-45ee-9c90-0b1e34a71340"). InnerVolumeSpecName "kube-api-access-x5ln8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:40:39 crc kubenswrapper[4818]: I1203 06:40:39.442663 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9897d8d1-c815-45ee-9c90-0b1e34a71340-util" (OuterVolumeSpecName: "util") pod "9897d8d1-c815-45ee-9c90-0b1e34a71340" (UID: "9897d8d1-c815-45ee-9c90-0b1e34a71340"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:40:39 crc kubenswrapper[4818]: I1203 06:40:39.511070 4818 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9897d8d1-c815-45ee-9c90-0b1e34a71340-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:40:39 crc kubenswrapper[4818]: I1203 06:40:39.511134 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5ln8\" (UniqueName: \"kubernetes.io/projected/9897d8d1-c815-45ee-9c90-0b1e34a71340-kube-api-access-x5ln8\") on node \"crc\" DevicePath \"\"" Dec 03 06:40:39 crc kubenswrapper[4818]: I1203 06:40:39.511154 4818 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9897d8d1-c815-45ee-9c90-0b1e34a71340-util\") on node \"crc\" DevicePath \"\"" Dec 03 06:40:39 crc kubenswrapper[4818]: I1203 06:40:39.782459 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" event={"ID":"9897d8d1-c815-45ee-9c90-0b1e34a71340","Type":"ContainerDied","Data":"d408b95d7d3c42882853ae9db16d1e70659dc904f7539b3de67851c18f36d6f5"} Dec 03 06:40:39 crc kubenswrapper[4818]: I1203 06:40:39.782514 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d408b95d7d3c42882853ae9db16d1e70659dc904f7539b3de67851c18f36d6f5" Dec 03 06:40:39 crc kubenswrapper[4818]: I1203 06:40:39.782993 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm" Dec 03 06:40:39 crc kubenswrapper[4818]: I1203 06:40:39.785803 4818 generic.go:334] "Generic (PLEG): container finished" podID="1df7f10a-718a-45d5-aad1-72fb5871bea8" containerID="541ec01d3b4a4c50f002dfff30455d5bd209b36f1f116fed6c171ce42abd62a4" exitCode=0 Dec 03 06:40:39 crc kubenswrapper[4818]: I1203 06:40:39.785869 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brv8j" event={"ID":"1df7f10a-718a-45d5-aad1-72fb5871bea8","Type":"ContainerDied","Data":"541ec01d3b4a4c50f002dfff30455d5bd209b36f1f116fed6c171ce42abd62a4"} Dec 03 06:40:40 crc kubenswrapper[4818]: I1203 06:40:40.796084 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brv8j" event={"ID":"1df7f10a-718a-45d5-aad1-72fb5871bea8","Type":"ContainerStarted","Data":"54178904fd0280491bebe50dcd35ff1472e5e8bc675ee3709fe3fde58f08052a"} Dec 03 06:40:40 crc kubenswrapper[4818]: I1203 06:40:40.816608 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-brv8j" podStartSLOduration=2.197960317 podStartE2EDuration="5.816584271s" podCreationTimestamp="2025-12-03 06:40:35 +0000 UTC" firstStartedPulling="2025-12-03 06:40:36.74615658 +0000 UTC m=+794.437765362" lastFinishedPulling="2025-12-03 06:40:40.364780524 +0000 UTC m=+798.056389316" observedRunningTime="2025-12-03 06:40:40.815603517 +0000 UTC m=+798.507212269" watchObservedRunningTime="2025-12-03 06:40:40.816584271 +0000 UTC m=+798.508193043" Dec 03 06:40:43 crc kubenswrapper[4818]: I1203 06:40:43.302178 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:40:43 crc kubenswrapper[4818]: I1203 06:40:43.302555 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:40:43 crc kubenswrapper[4818]: I1203 06:40:43.302599 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:40:43 crc kubenswrapper[4818]: I1203 06:40:43.303230 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d9ea0215376ae671f51a5593609153a4204668620744066ed3e3215741cd9e5f"} pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 06:40:43 crc kubenswrapper[4818]: I1203 06:40:43.303287 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" containerID="cri-o://d9ea0215376ae671f51a5593609153a4204668620744066ed3e3215741cd9e5f" gracePeriod=600 Dec 03 06:40:44 crc kubenswrapper[4818]: I1203 06:40:44.825887 4818 generic.go:334] "Generic (PLEG): container finished" podID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerID="d9ea0215376ae671f51a5593609153a4204668620744066ed3e3215741cd9e5f" exitCode=0 Dec 03 06:40:44 crc kubenswrapper[4818]: I1203 06:40:44.826241 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerDied","Data":"d9ea0215376ae671f51a5593609153a4204668620744066ed3e3215741cd9e5f"} Dec 03 06:40:44 crc kubenswrapper[4818]: I1203 06:40:44.826278 4818 scope.go:117] "RemoveContainer" containerID="29214294c382d4bed0dc9a820848900ea80c550461b1b157cc6bdd05a78333e5" Dec 03 06:40:45 crc kubenswrapper[4818]: I1203 06:40:45.552412 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-brv8j" Dec 03 06:40:45 crc kubenswrapper[4818]: I1203 06:40:45.552834 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-brv8j" Dec 03 06:40:45 crc kubenswrapper[4818]: I1203 06:40:45.848434 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"9e30bff309bf8bc13ed470ed28234141d81432e279f1c0a4251c4793b659775c"} Dec 03 06:40:46 crc kubenswrapper[4818]: I1203 06:40:46.655992 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-brv8j" podUID="1df7f10a-718a-45d5-aad1-72fb5871bea8" containerName="registry-server" probeResult="failure" output=< Dec 03 06:40:46 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Dec 03 06:40:46 crc kubenswrapper[4818]: > Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.082252 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk"] Dec 03 06:40:49 crc kubenswrapper[4818]: E1203 06:40:49.082618 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9897d8d1-c815-45ee-9c90-0b1e34a71340" containerName="pull" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.082631 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9897d8d1-c815-45ee-9c90-0b1e34a71340" containerName="pull" Dec 03 06:40:49 crc kubenswrapper[4818]: E1203 06:40:49.082652 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9897d8d1-c815-45ee-9c90-0b1e34a71340" containerName="extract" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.082658 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9897d8d1-c815-45ee-9c90-0b1e34a71340" containerName="extract" Dec 03 06:40:49 crc kubenswrapper[4818]: E1203 06:40:49.082673 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9897d8d1-c815-45ee-9c90-0b1e34a71340" containerName="util" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.082681 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9897d8d1-c815-45ee-9c90-0b1e34a71340" containerName="util" Dec 03 06:40:49 crc kubenswrapper[4818]: E1203 06:40:49.082695 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d2839d4-e624-4d61-9225-894a998e9e9e" containerName="console" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.082701 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d2839d4-e624-4d61-9225-894a998e9e9e" containerName="console" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.082917 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="9897d8d1-c815-45ee-9c90-0b1e34a71340" containerName="extract" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.082939 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d2839d4-e624-4d61-9225-894a998e9e9e" containerName="console" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.083463 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.090506 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.090764 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.091056 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.091297 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.092185 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-cwbvr" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.097730 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk"] Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.242383 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d227f431-7cf6-43ed-85d0-6e44812df281-apiservice-cert\") pod \"metallb-operator-controller-manager-5545bb6cbd-gjkwk\" (UID: \"d227f431-7cf6-43ed-85d0-6e44812df281\") " pod="metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.242441 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d227f431-7cf6-43ed-85d0-6e44812df281-webhook-cert\") pod \"metallb-operator-controller-manager-5545bb6cbd-gjkwk\" (UID: \"d227f431-7cf6-43ed-85d0-6e44812df281\") " pod="metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.242481 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c52w\" (UniqueName: \"kubernetes.io/projected/d227f431-7cf6-43ed-85d0-6e44812df281-kube-api-access-9c52w\") pod \"metallb-operator-controller-manager-5545bb6cbd-gjkwk\" (UID: \"d227f431-7cf6-43ed-85d0-6e44812df281\") " pod="metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.328325 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc"] Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.328999 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.330624 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.330946 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.332138 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-q4474" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.344078 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc"] Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.344977 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d227f431-7cf6-43ed-85d0-6e44812df281-apiservice-cert\") pod \"metallb-operator-controller-manager-5545bb6cbd-gjkwk\" (UID: \"d227f431-7cf6-43ed-85d0-6e44812df281\") " pod="metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.345034 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d227f431-7cf6-43ed-85d0-6e44812df281-webhook-cert\") pod \"metallb-operator-controller-manager-5545bb6cbd-gjkwk\" (UID: \"d227f431-7cf6-43ed-85d0-6e44812df281\") " pod="metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.345074 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c52w\" (UniqueName: \"kubernetes.io/projected/d227f431-7cf6-43ed-85d0-6e44812df281-kube-api-access-9c52w\") pod \"metallb-operator-controller-manager-5545bb6cbd-gjkwk\" (UID: \"d227f431-7cf6-43ed-85d0-6e44812df281\") " pod="metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.345109 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhg8z\" (UniqueName: \"kubernetes.io/projected/4563221c-8432-4038-b0e6-fdbac56b99e0-kube-api-access-xhg8z\") pod \"metallb-operator-webhook-server-56f4c8658f-s8ghc\" (UID: \"4563221c-8432-4038-b0e6-fdbac56b99e0\") " pod="metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.345150 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4563221c-8432-4038-b0e6-fdbac56b99e0-apiservice-cert\") pod \"metallb-operator-webhook-server-56f4c8658f-s8ghc\" (UID: \"4563221c-8432-4038-b0e6-fdbac56b99e0\") " pod="metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.345179 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4563221c-8432-4038-b0e6-fdbac56b99e0-webhook-cert\") pod \"metallb-operator-webhook-server-56f4c8658f-s8ghc\" (UID: \"4563221c-8432-4038-b0e6-fdbac56b99e0\") " pod="metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.351369 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d227f431-7cf6-43ed-85d0-6e44812df281-webhook-cert\") pod \"metallb-operator-controller-manager-5545bb6cbd-gjkwk\" (UID: \"d227f431-7cf6-43ed-85d0-6e44812df281\") " pod="metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.352625 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d227f431-7cf6-43ed-85d0-6e44812df281-apiservice-cert\") pod \"metallb-operator-controller-manager-5545bb6cbd-gjkwk\" (UID: \"d227f431-7cf6-43ed-85d0-6e44812df281\") " pod="metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.372692 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c52w\" (UniqueName: \"kubernetes.io/projected/d227f431-7cf6-43ed-85d0-6e44812df281-kube-api-access-9c52w\") pod \"metallb-operator-controller-manager-5545bb6cbd-gjkwk\" (UID: \"d227f431-7cf6-43ed-85d0-6e44812df281\") " pod="metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.428496 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.446575 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4563221c-8432-4038-b0e6-fdbac56b99e0-webhook-cert\") pod \"metallb-operator-webhook-server-56f4c8658f-s8ghc\" (UID: \"4563221c-8432-4038-b0e6-fdbac56b99e0\") " pod="metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.447261 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhg8z\" (UniqueName: \"kubernetes.io/projected/4563221c-8432-4038-b0e6-fdbac56b99e0-kube-api-access-xhg8z\") pod \"metallb-operator-webhook-server-56f4c8658f-s8ghc\" (UID: \"4563221c-8432-4038-b0e6-fdbac56b99e0\") " pod="metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.447616 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4563221c-8432-4038-b0e6-fdbac56b99e0-apiservice-cert\") pod \"metallb-operator-webhook-server-56f4c8658f-s8ghc\" (UID: \"4563221c-8432-4038-b0e6-fdbac56b99e0\") " pod="metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.450407 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4563221c-8432-4038-b0e6-fdbac56b99e0-webhook-cert\") pod \"metallb-operator-webhook-server-56f4c8658f-s8ghc\" (UID: \"4563221c-8432-4038-b0e6-fdbac56b99e0\") " pod="metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.451022 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4563221c-8432-4038-b0e6-fdbac56b99e0-apiservice-cert\") pod \"metallb-operator-webhook-server-56f4c8658f-s8ghc\" (UID: \"4563221c-8432-4038-b0e6-fdbac56b99e0\") " pod="metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.470623 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhg8z\" (UniqueName: \"kubernetes.io/projected/4563221c-8432-4038-b0e6-fdbac56b99e0-kube-api-access-xhg8z\") pod \"metallb-operator-webhook-server-56f4c8658f-s8ghc\" (UID: \"4563221c-8432-4038-b0e6-fdbac56b99e0\") " pod="metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.650056 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc" Dec 03 06:40:49 crc kubenswrapper[4818]: I1203 06:40:49.968262 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk"] Dec 03 06:40:49 crc kubenswrapper[4818]: W1203 06:40:49.973647 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd227f431_7cf6_43ed_85d0_6e44812df281.slice/crio-54689d6cbd16e0f6c37c2e50bf5f3615c2e7766974a4ca5c388a1ea3df0ffd00 WatchSource:0}: Error finding container 54689d6cbd16e0f6c37c2e50bf5f3615c2e7766974a4ca5c388a1ea3df0ffd00: Status 404 returned error can't find the container with id 54689d6cbd16e0f6c37c2e50bf5f3615c2e7766974a4ca5c388a1ea3df0ffd00 Dec 03 06:40:50 crc kubenswrapper[4818]: I1203 06:40:50.183768 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc"] Dec 03 06:40:50 crc kubenswrapper[4818]: W1203 06:40:50.191234 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4563221c_8432_4038_b0e6_fdbac56b99e0.slice/crio-7813cb8438665ef737f8abe9a4dd18b317c60f45ba7d64e6705b66b99a4dedb8 WatchSource:0}: Error finding container 7813cb8438665ef737f8abe9a4dd18b317c60f45ba7d64e6705b66b99a4dedb8: Status 404 returned error can't find the container with id 7813cb8438665ef737f8abe9a4dd18b317c60f45ba7d64e6705b66b99a4dedb8 Dec 03 06:40:50 crc kubenswrapper[4818]: I1203 06:40:50.876700 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc" event={"ID":"4563221c-8432-4038-b0e6-fdbac56b99e0","Type":"ContainerStarted","Data":"7813cb8438665ef737f8abe9a4dd18b317c60f45ba7d64e6705b66b99a4dedb8"} Dec 03 06:40:50 crc kubenswrapper[4818]: I1203 06:40:50.878158 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk" event={"ID":"d227f431-7cf6-43ed-85d0-6e44812df281","Type":"ContainerStarted","Data":"54689d6cbd16e0f6c37c2e50bf5f3615c2e7766974a4ca5c388a1ea3df0ffd00"} Dec 03 06:40:53 crc kubenswrapper[4818]: I1203 06:40:53.894836 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk" event={"ID":"d227f431-7cf6-43ed-85d0-6e44812df281","Type":"ContainerStarted","Data":"6685d59b415d2e49457c471d233fa768a6cb021b7963937ccfccd0206e119189"} Dec 03 06:40:53 crc kubenswrapper[4818]: I1203 06:40:53.896103 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk" Dec 03 06:40:53 crc kubenswrapper[4818]: I1203 06:40:53.911846 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk" podStartSLOduration=1.516399818 podStartE2EDuration="4.911808011s" podCreationTimestamp="2025-12-03 06:40:49 +0000 UTC" firstStartedPulling="2025-12-03 06:40:49.976528127 +0000 UTC m=+807.668136879" lastFinishedPulling="2025-12-03 06:40:53.37193632 +0000 UTC m=+811.063545072" observedRunningTime="2025-12-03 06:40:53.910335044 +0000 UTC m=+811.601943796" watchObservedRunningTime="2025-12-03 06:40:53.911808011 +0000 UTC m=+811.603416763" Dec 03 06:40:55 crc kubenswrapper[4818]: I1203 06:40:55.610585 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-brv8j" Dec 03 06:40:55 crc kubenswrapper[4818]: I1203 06:40:55.672491 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-brv8j" Dec 03 06:40:56 crc kubenswrapper[4818]: I1203 06:40:56.396830 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-brv8j"] Dec 03 06:40:56 crc kubenswrapper[4818]: I1203 06:40:56.921510 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-brv8j" podUID="1df7f10a-718a-45d5-aad1-72fb5871bea8" containerName="registry-server" containerID="cri-o://54178904fd0280491bebe50dcd35ff1472e5e8bc675ee3709fe3fde58f08052a" gracePeriod=2 Dec 03 06:40:56 crc kubenswrapper[4818]: I1203 06:40:56.922319 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc" event={"ID":"4563221c-8432-4038-b0e6-fdbac56b99e0","Type":"ContainerStarted","Data":"7251af154ebb48fdcbc9353aecf59a512e2227683173d0c51770d3e7e3156df1"} Dec 03 06:40:56 crc kubenswrapper[4818]: I1203 06:40:56.922515 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc" Dec 03 06:40:56 crc kubenswrapper[4818]: I1203 06:40:56.950370 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc" podStartSLOduration=1.782790959 podStartE2EDuration="7.950354529s" podCreationTimestamp="2025-12-03 06:40:49 +0000 UTC" firstStartedPulling="2025-12-03 06:40:50.198746314 +0000 UTC m=+807.890355066" lastFinishedPulling="2025-12-03 06:40:56.366309874 +0000 UTC m=+814.057918636" observedRunningTime="2025-12-03 06:40:56.949875797 +0000 UTC m=+814.641484549" watchObservedRunningTime="2025-12-03 06:40:56.950354529 +0000 UTC m=+814.641963281" Dec 03 06:40:57 crc kubenswrapper[4818]: I1203 06:40:57.266754 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-brv8j" Dec 03 06:40:57 crc kubenswrapper[4818]: I1203 06:40:57.338963 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df7f10a-718a-45d5-aad1-72fb5871bea8-catalog-content\") pod \"1df7f10a-718a-45d5-aad1-72fb5871bea8\" (UID: \"1df7f10a-718a-45d5-aad1-72fb5871bea8\") " Dec 03 06:40:57 crc kubenswrapper[4818]: I1203 06:40:57.339016 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr6fv\" (UniqueName: \"kubernetes.io/projected/1df7f10a-718a-45d5-aad1-72fb5871bea8-kube-api-access-qr6fv\") pod \"1df7f10a-718a-45d5-aad1-72fb5871bea8\" (UID: \"1df7f10a-718a-45d5-aad1-72fb5871bea8\") " Dec 03 06:40:57 crc kubenswrapper[4818]: I1203 06:40:57.339052 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df7f10a-718a-45d5-aad1-72fb5871bea8-utilities\") pod \"1df7f10a-718a-45d5-aad1-72fb5871bea8\" (UID: \"1df7f10a-718a-45d5-aad1-72fb5871bea8\") " Dec 03 06:40:57 crc kubenswrapper[4818]: I1203 06:40:57.340128 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1df7f10a-718a-45d5-aad1-72fb5871bea8-utilities" (OuterVolumeSpecName: "utilities") pod "1df7f10a-718a-45d5-aad1-72fb5871bea8" (UID: "1df7f10a-718a-45d5-aad1-72fb5871bea8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:40:57 crc kubenswrapper[4818]: I1203 06:40:57.347130 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1df7f10a-718a-45d5-aad1-72fb5871bea8-kube-api-access-qr6fv" (OuterVolumeSpecName: "kube-api-access-qr6fv") pod "1df7f10a-718a-45d5-aad1-72fb5871bea8" (UID: "1df7f10a-718a-45d5-aad1-72fb5871bea8"). InnerVolumeSpecName "kube-api-access-qr6fv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:40:57 crc kubenswrapper[4818]: I1203 06:40:57.440202 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df7f10a-718a-45d5-aad1-72fb5871bea8-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:40:57 crc kubenswrapper[4818]: I1203 06:40:57.440238 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr6fv\" (UniqueName: \"kubernetes.io/projected/1df7f10a-718a-45d5-aad1-72fb5871bea8-kube-api-access-qr6fv\") on node \"crc\" DevicePath \"\"" Dec 03 06:40:57 crc kubenswrapper[4818]: I1203 06:40:57.491654 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1df7f10a-718a-45d5-aad1-72fb5871bea8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1df7f10a-718a-45d5-aad1-72fb5871bea8" (UID: "1df7f10a-718a-45d5-aad1-72fb5871bea8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:40:57 crc kubenswrapper[4818]: I1203 06:40:57.541386 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df7f10a-718a-45d5-aad1-72fb5871bea8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:40:57 crc kubenswrapper[4818]: I1203 06:40:57.931426 4818 generic.go:334] "Generic (PLEG): container finished" podID="1df7f10a-718a-45d5-aad1-72fb5871bea8" containerID="54178904fd0280491bebe50dcd35ff1472e5e8bc675ee3709fe3fde58f08052a" exitCode=0 Dec 03 06:40:57 crc kubenswrapper[4818]: I1203 06:40:57.931601 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brv8j" event={"ID":"1df7f10a-718a-45d5-aad1-72fb5871bea8","Type":"ContainerDied","Data":"54178904fd0280491bebe50dcd35ff1472e5e8bc675ee3709fe3fde58f08052a"} Dec 03 06:40:57 crc kubenswrapper[4818]: I1203 06:40:57.931624 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-brv8j" Dec 03 06:40:57 crc kubenswrapper[4818]: I1203 06:40:57.931692 4818 scope.go:117] "RemoveContainer" containerID="54178904fd0280491bebe50dcd35ff1472e5e8bc675ee3709fe3fde58f08052a" Dec 03 06:40:57 crc kubenswrapper[4818]: I1203 06:40:57.931671 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brv8j" event={"ID":"1df7f10a-718a-45d5-aad1-72fb5871bea8","Type":"ContainerDied","Data":"f604d2396d4833b774661e06fea8a486b13b784511debe336e1d704a9f397033"} Dec 03 06:40:57 crc kubenswrapper[4818]: I1203 06:40:57.959094 4818 scope.go:117] "RemoveContainer" containerID="541ec01d3b4a4c50f002dfff30455d5bd209b36f1f116fed6c171ce42abd62a4" Dec 03 06:40:57 crc kubenswrapper[4818]: I1203 06:40:57.987869 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-brv8j"] Dec 03 06:40:58 crc kubenswrapper[4818]: I1203 06:40:58.001666 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-brv8j"] Dec 03 06:40:58 crc kubenswrapper[4818]: I1203 06:40:58.029202 4818 scope.go:117] "RemoveContainer" containerID="2852a7705fbaa7e5b8d7216abb77d4d6e2995a9ec18394a97fed1b193341a7bd" Dec 03 06:40:58 crc kubenswrapper[4818]: I1203 06:40:58.054317 4818 scope.go:117] "RemoveContainer" containerID="54178904fd0280491bebe50dcd35ff1472e5e8bc675ee3709fe3fde58f08052a" Dec 03 06:40:58 crc kubenswrapper[4818]: E1203 06:40:58.055015 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54178904fd0280491bebe50dcd35ff1472e5e8bc675ee3709fe3fde58f08052a\": container with ID starting with 54178904fd0280491bebe50dcd35ff1472e5e8bc675ee3709fe3fde58f08052a not found: ID does not exist" containerID="54178904fd0280491bebe50dcd35ff1472e5e8bc675ee3709fe3fde58f08052a" Dec 03 06:40:58 crc kubenswrapper[4818]: I1203 06:40:58.055129 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54178904fd0280491bebe50dcd35ff1472e5e8bc675ee3709fe3fde58f08052a"} err="failed to get container status \"54178904fd0280491bebe50dcd35ff1472e5e8bc675ee3709fe3fde58f08052a\": rpc error: code = NotFound desc = could not find container \"54178904fd0280491bebe50dcd35ff1472e5e8bc675ee3709fe3fde58f08052a\": container with ID starting with 54178904fd0280491bebe50dcd35ff1472e5e8bc675ee3709fe3fde58f08052a not found: ID does not exist" Dec 03 06:40:58 crc kubenswrapper[4818]: I1203 06:40:58.055190 4818 scope.go:117] "RemoveContainer" containerID="541ec01d3b4a4c50f002dfff30455d5bd209b36f1f116fed6c171ce42abd62a4" Dec 03 06:40:58 crc kubenswrapper[4818]: E1203 06:40:58.057450 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"541ec01d3b4a4c50f002dfff30455d5bd209b36f1f116fed6c171ce42abd62a4\": container with ID starting with 541ec01d3b4a4c50f002dfff30455d5bd209b36f1f116fed6c171ce42abd62a4 not found: ID does not exist" containerID="541ec01d3b4a4c50f002dfff30455d5bd209b36f1f116fed6c171ce42abd62a4" Dec 03 06:40:58 crc kubenswrapper[4818]: I1203 06:40:58.057505 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"541ec01d3b4a4c50f002dfff30455d5bd209b36f1f116fed6c171ce42abd62a4"} err="failed to get container status \"541ec01d3b4a4c50f002dfff30455d5bd209b36f1f116fed6c171ce42abd62a4\": rpc error: code = NotFound desc = could not find container \"541ec01d3b4a4c50f002dfff30455d5bd209b36f1f116fed6c171ce42abd62a4\": container with ID starting with 541ec01d3b4a4c50f002dfff30455d5bd209b36f1f116fed6c171ce42abd62a4 not found: ID does not exist" Dec 03 06:40:58 crc kubenswrapper[4818]: I1203 06:40:58.057535 4818 scope.go:117] "RemoveContainer" containerID="2852a7705fbaa7e5b8d7216abb77d4d6e2995a9ec18394a97fed1b193341a7bd" Dec 03 06:40:58 crc kubenswrapper[4818]: E1203 06:40:58.058035 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2852a7705fbaa7e5b8d7216abb77d4d6e2995a9ec18394a97fed1b193341a7bd\": container with ID starting with 2852a7705fbaa7e5b8d7216abb77d4d6e2995a9ec18394a97fed1b193341a7bd not found: ID does not exist" containerID="2852a7705fbaa7e5b8d7216abb77d4d6e2995a9ec18394a97fed1b193341a7bd" Dec 03 06:40:58 crc kubenswrapper[4818]: I1203 06:40:58.058075 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2852a7705fbaa7e5b8d7216abb77d4d6e2995a9ec18394a97fed1b193341a7bd"} err="failed to get container status \"2852a7705fbaa7e5b8d7216abb77d4d6e2995a9ec18394a97fed1b193341a7bd\": rpc error: code = NotFound desc = could not find container \"2852a7705fbaa7e5b8d7216abb77d4d6e2995a9ec18394a97fed1b193341a7bd\": container with ID starting with 2852a7705fbaa7e5b8d7216abb77d4d6e2995a9ec18394a97fed1b193341a7bd not found: ID does not exist" Dec 03 06:40:58 crc kubenswrapper[4818]: I1203 06:40:58.745259 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1df7f10a-718a-45d5-aad1-72fb5871bea8" path="/var/lib/kubelet/pods/1df7f10a-718a-45d5-aad1-72fb5871bea8/volumes" Dec 03 06:41:09 crc kubenswrapper[4818]: I1203 06:41:09.655376 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-56f4c8658f-s8ghc" Dec 03 06:41:29 crc kubenswrapper[4818]: I1203 06:41:29.432442 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5545bb6cbd-gjkwk" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.224761 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-tkg7w"] Dec 03 06:41:30 crc kubenswrapper[4818]: E1203 06:41:30.225113 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df7f10a-718a-45d5-aad1-72fb5871bea8" containerName="registry-server" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.225134 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df7f10a-718a-45d5-aad1-72fb5871bea8" containerName="registry-server" Dec 03 06:41:30 crc kubenswrapper[4818]: E1203 06:41:30.225155 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df7f10a-718a-45d5-aad1-72fb5871bea8" containerName="extract-content" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.225165 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df7f10a-718a-45d5-aad1-72fb5871bea8" containerName="extract-content" Dec 03 06:41:30 crc kubenswrapper[4818]: E1203 06:41:30.225176 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df7f10a-718a-45d5-aad1-72fb5871bea8" containerName="extract-utilities" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.225187 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df7f10a-718a-45d5-aad1-72fb5871bea8" containerName="extract-utilities" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.225310 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df7f10a-718a-45d5-aad1-72fb5871bea8" containerName="registry-server" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.225810 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-tkg7w" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.227766 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-5577l" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.227776 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.230036 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-hwmtn"] Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.233030 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.234462 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.241286 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-tkg7w"] Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.241451 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.326332 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-4m9n7"] Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.327144 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-4m9n7" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.329535 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.330102 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.330282 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.330436 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-gz74n" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.364235 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-7nr97"] Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.366229 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-7nr97" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.368316 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.370031 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0a528d4d-3d84-40c2-a4b3-207001ba3579-frr-startup\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.370079 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0a528d4d-3d84-40c2-a4b3-207001ba3579-metrics\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.370108 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt8zd\" (UniqueName: \"kubernetes.io/projected/0a528d4d-3d84-40c2-a4b3-207001ba3579-kube-api-access-kt8zd\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.370178 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gbw6\" (UniqueName: \"kubernetes.io/projected/f55ffc30-37c1-45e8-8e63-c447ca4d6e96-kube-api-access-5gbw6\") pod \"frr-k8s-webhook-server-7fcb986d4-tkg7w\" (UID: \"f55ffc30-37c1-45e8-8e63-c447ca4d6e96\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-tkg7w" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.370209 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a528d4d-3d84-40c2-a4b3-207001ba3579-metrics-certs\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.370235 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f55ffc30-37c1-45e8-8e63-c447ca4d6e96-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-tkg7w\" (UID: \"f55ffc30-37c1-45e8-8e63-c447ca4d6e96\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-tkg7w" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.370275 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0a528d4d-3d84-40c2-a4b3-207001ba3579-frr-sockets\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.370347 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0a528d4d-3d84-40c2-a4b3-207001ba3579-reloader\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.370383 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0a528d4d-3d84-40c2-a4b3-207001ba3579-frr-conf\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.378063 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-7nr97"] Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.471320 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gbw6\" (UniqueName: \"kubernetes.io/projected/f55ffc30-37c1-45e8-8e63-c447ca4d6e96-kube-api-access-5gbw6\") pod \"frr-k8s-webhook-server-7fcb986d4-tkg7w\" (UID: \"f55ffc30-37c1-45e8-8e63-c447ca4d6e96\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-tkg7w" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.471364 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a528d4d-3d84-40c2-a4b3-207001ba3579-metrics-certs\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.471392 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49h95\" (UniqueName: \"kubernetes.io/projected/bf628f10-0970-4651-8499-5302e78c8d6d-kube-api-access-49h95\") pod \"controller-f8648f98b-7nr97\" (UID: \"bf628f10-0970-4651-8499-5302e78c8d6d\") " pod="metallb-system/controller-f8648f98b-7nr97" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.471411 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f55ffc30-37c1-45e8-8e63-c447ca4d6e96-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-tkg7w\" (UID: \"f55ffc30-37c1-45e8-8e63-c447ca4d6e96\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-tkg7w" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.471431 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/860c17a3-a17f-4ac3-9310-c633df611f6d-metallb-excludel2\") pod \"speaker-4m9n7\" (UID: \"860c17a3-a17f-4ac3-9310-c633df611f6d\") " pod="metallb-system/speaker-4m9n7" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.471472 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/860c17a3-a17f-4ac3-9310-c633df611f6d-metrics-certs\") pod \"speaker-4m9n7\" (UID: \"860c17a3-a17f-4ac3-9310-c633df611f6d\") " pod="metallb-system/speaker-4m9n7" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.471596 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wccxz\" (UniqueName: \"kubernetes.io/projected/860c17a3-a17f-4ac3-9310-c633df611f6d-kube-api-access-wccxz\") pod \"speaker-4m9n7\" (UID: \"860c17a3-a17f-4ac3-9310-c633df611f6d\") " pod="metallb-system/speaker-4m9n7" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.471618 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bf628f10-0970-4651-8499-5302e78c8d6d-metrics-certs\") pod \"controller-f8648f98b-7nr97\" (UID: \"bf628f10-0970-4651-8499-5302e78c8d6d\") " pod="metallb-system/controller-f8648f98b-7nr97" Dec 03 06:41:30 crc kubenswrapper[4818]: E1203 06:41:30.471773 4818 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.471794 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0a528d4d-3d84-40c2-a4b3-207001ba3579-frr-sockets\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.471881 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0a528d4d-3d84-40c2-a4b3-207001ba3579-reloader\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: E1203 06:41:30.471934 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f55ffc30-37c1-45e8-8e63-c447ca4d6e96-cert podName:f55ffc30-37c1-45e8-8e63-c447ca4d6e96 nodeName:}" failed. No retries permitted until 2025-12-03 06:41:30.971884175 +0000 UTC m=+848.663492927 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f55ffc30-37c1-45e8-8e63-c447ca4d6e96-cert") pod "frr-k8s-webhook-server-7fcb986d4-tkg7w" (UID: "f55ffc30-37c1-45e8-8e63-c447ca4d6e96") : secret "frr-k8s-webhook-server-cert" not found Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.472008 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/860c17a3-a17f-4ac3-9310-c633df611f6d-memberlist\") pod \"speaker-4m9n7\" (UID: \"860c17a3-a17f-4ac3-9310-c633df611f6d\") " pod="metallb-system/speaker-4m9n7" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.472114 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0a528d4d-3d84-40c2-a4b3-207001ba3579-frr-conf\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.472297 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0a528d4d-3d84-40c2-a4b3-207001ba3579-frr-startup\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.472356 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0a528d4d-3d84-40c2-a4b3-207001ba3579-metrics\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.472400 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt8zd\" (UniqueName: \"kubernetes.io/projected/0a528d4d-3d84-40c2-a4b3-207001ba3579-kube-api-access-kt8zd\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.472452 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf628f10-0970-4651-8499-5302e78c8d6d-cert\") pod \"controller-f8648f98b-7nr97\" (UID: \"bf628f10-0970-4651-8499-5302e78c8d6d\") " pod="metallb-system/controller-f8648f98b-7nr97" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.474503 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/0a528d4d-3d84-40c2-a4b3-207001ba3579-reloader\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.474613 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/0a528d4d-3d84-40c2-a4b3-207001ba3579-frr-conf\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.474673 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/0a528d4d-3d84-40c2-a4b3-207001ba3579-frr-sockets\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.474810 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/0a528d4d-3d84-40c2-a4b3-207001ba3579-metrics\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.475092 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/0a528d4d-3d84-40c2-a4b3-207001ba3579-frr-startup\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.479424 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a528d4d-3d84-40c2-a4b3-207001ba3579-metrics-certs\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.495883 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt8zd\" (UniqueName: \"kubernetes.io/projected/0a528d4d-3d84-40c2-a4b3-207001ba3579-kube-api-access-kt8zd\") pod \"frr-k8s-hwmtn\" (UID: \"0a528d4d-3d84-40c2-a4b3-207001ba3579\") " pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.496060 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gbw6\" (UniqueName: \"kubernetes.io/projected/f55ffc30-37c1-45e8-8e63-c447ca4d6e96-kube-api-access-5gbw6\") pod \"frr-k8s-webhook-server-7fcb986d4-tkg7w\" (UID: \"f55ffc30-37c1-45e8-8e63-c447ca4d6e96\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-tkg7w" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.557791 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.574146 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49h95\" (UniqueName: \"kubernetes.io/projected/bf628f10-0970-4651-8499-5302e78c8d6d-kube-api-access-49h95\") pod \"controller-f8648f98b-7nr97\" (UID: \"bf628f10-0970-4651-8499-5302e78c8d6d\") " pod="metallb-system/controller-f8648f98b-7nr97" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.574476 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/860c17a3-a17f-4ac3-9310-c633df611f6d-metallb-excludel2\") pod \"speaker-4m9n7\" (UID: \"860c17a3-a17f-4ac3-9310-c633df611f6d\") " pod="metallb-system/speaker-4m9n7" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.574505 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/860c17a3-a17f-4ac3-9310-c633df611f6d-metrics-certs\") pod \"speaker-4m9n7\" (UID: \"860c17a3-a17f-4ac3-9310-c633df611f6d\") " pod="metallb-system/speaker-4m9n7" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.574529 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wccxz\" (UniqueName: \"kubernetes.io/projected/860c17a3-a17f-4ac3-9310-c633df611f6d-kube-api-access-wccxz\") pod \"speaker-4m9n7\" (UID: \"860c17a3-a17f-4ac3-9310-c633df611f6d\") " pod="metallb-system/speaker-4m9n7" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.574552 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bf628f10-0970-4651-8499-5302e78c8d6d-metrics-certs\") pod \"controller-f8648f98b-7nr97\" (UID: \"bf628f10-0970-4651-8499-5302e78c8d6d\") " pod="metallb-system/controller-f8648f98b-7nr97" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.574586 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/860c17a3-a17f-4ac3-9310-c633df611f6d-memberlist\") pod \"speaker-4m9n7\" (UID: \"860c17a3-a17f-4ac3-9310-c633df611f6d\") " pod="metallb-system/speaker-4m9n7" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.574666 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf628f10-0970-4651-8499-5302e78c8d6d-cert\") pod \"controller-f8648f98b-7nr97\" (UID: \"bf628f10-0970-4651-8499-5302e78c8d6d\") " pod="metallb-system/controller-f8648f98b-7nr97" Dec 03 06:41:30 crc kubenswrapper[4818]: E1203 06:41:30.574763 4818 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 06:41:30 crc kubenswrapper[4818]: E1203 06:41:30.574847 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/860c17a3-a17f-4ac3-9310-c633df611f6d-memberlist podName:860c17a3-a17f-4ac3-9310-c633df611f6d nodeName:}" failed. No retries permitted until 2025-12-03 06:41:31.074826296 +0000 UTC m=+848.766435048 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/860c17a3-a17f-4ac3-9310-c633df611f6d-memberlist") pod "speaker-4m9n7" (UID: "860c17a3-a17f-4ac3-9310-c633df611f6d") : secret "metallb-memberlist" not found Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.575613 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/860c17a3-a17f-4ac3-9310-c633df611f6d-metallb-excludel2\") pod \"speaker-4m9n7\" (UID: \"860c17a3-a17f-4ac3-9310-c633df611f6d\") " pod="metallb-system/speaker-4m9n7" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.579887 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bf628f10-0970-4651-8499-5302e78c8d6d-metrics-certs\") pod \"controller-f8648f98b-7nr97\" (UID: \"bf628f10-0970-4651-8499-5302e78c8d6d\") " pod="metallb-system/controller-f8648f98b-7nr97" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.581206 4818 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.581422 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/860c17a3-a17f-4ac3-9310-c633df611f6d-metrics-certs\") pod \"speaker-4m9n7\" (UID: \"860c17a3-a17f-4ac3-9310-c633df611f6d\") " pod="metallb-system/speaker-4m9n7" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.596431 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wccxz\" (UniqueName: \"kubernetes.io/projected/860c17a3-a17f-4ac3-9310-c633df611f6d-kube-api-access-wccxz\") pod \"speaker-4m9n7\" (UID: \"860c17a3-a17f-4ac3-9310-c633df611f6d\") " pod="metallb-system/speaker-4m9n7" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.597533 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49h95\" (UniqueName: \"kubernetes.io/projected/bf628f10-0970-4651-8499-5302e78c8d6d-kube-api-access-49h95\") pod \"controller-f8648f98b-7nr97\" (UID: \"bf628f10-0970-4651-8499-5302e78c8d6d\") " pod="metallb-system/controller-f8648f98b-7nr97" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.600203 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf628f10-0970-4651-8499-5302e78c8d6d-cert\") pod \"controller-f8648f98b-7nr97\" (UID: \"bf628f10-0970-4651-8499-5302e78c8d6d\") " pod="metallb-system/controller-f8648f98b-7nr97" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.693195 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-7nr97" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.982546 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f55ffc30-37c1-45e8-8e63-c447ca4d6e96-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-tkg7w\" (UID: \"f55ffc30-37c1-45e8-8e63-c447ca4d6e96\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-tkg7w" Dec 03 06:41:30 crc kubenswrapper[4818]: I1203 06:41:30.990981 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f55ffc30-37c1-45e8-8e63-c447ca4d6e96-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-tkg7w\" (UID: \"f55ffc30-37c1-45e8-8e63-c447ca4d6e96\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-tkg7w" Dec 03 06:41:31 crc kubenswrapper[4818]: I1203 06:41:31.083979 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/860c17a3-a17f-4ac3-9310-c633df611f6d-memberlist\") pod \"speaker-4m9n7\" (UID: \"860c17a3-a17f-4ac3-9310-c633df611f6d\") " pod="metallb-system/speaker-4m9n7" Dec 03 06:41:31 crc kubenswrapper[4818]: E1203 06:41:31.084162 4818 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 06:41:31 crc kubenswrapper[4818]: E1203 06:41:31.084214 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/860c17a3-a17f-4ac3-9310-c633df611f6d-memberlist podName:860c17a3-a17f-4ac3-9310-c633df611f6d nodeName:}" failed. No retries permitted until 2025-12-03 06:41:32.08419859 +0000 UTC m=+849.775807342 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/860c17a3-a17f-4ac3-9310-c633df611f6d-memberlist") pod "speaker-4m9n7" (UID: "860c17a3-a17f-4ac3-9310-c633df611f6d") : secret "metallb-memberlist" not found Dec 03 06:41:31 crc kubenswrapper[4818]: I1203 06:41:31.091003 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-7nr97"] Dec 03 06:41:31 crc kubenswrapper[4818]: W1203 06:41:31.093669 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf628f10_0970_4651_8499_5302e78c8d6d.slice/crio-cf700b14320300b85b69ff17207fc7e92e3317c171757490551fc4d8f02c3acc WatchSource:0}: Error finding container cf700b14320300b85b69ff17207fc7e92e3317c171757490551fc4d8f02c3acc: Status 404 returned error can't find the container with id cf700b14320300b85b69ff17207fc7e92e3317c171757490551fc4d8f02c3acc Dec 03 06:41:31 crc kubenswrapper[4818]: I1203 06:41:31.143172 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hwmtn" event={"ID":"0a528d4d-3d84-40c2-a4b3-207001ba3579","Type":"ContainerStarted","Data":"f3f02d2b3911d60f1b1f079a9b015e46683c5ca9f405a5ac5a81831cb57cc0af"} Dec 03 06:41:31 crc kubenswrapper[4818]: I1203 06:41:31.144012 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-7nr97" event={"ID":"bf628f10-0970-4651-8499-5302e78c8d6d","Type":"ContainerStarted","Data":"cf700b14320300b85b69ff17207fc7e92e3317c171757490551fc4d8f02c3acc"} Dec 03 06:41:31 crc kubenswrapper[4818]: I1203 06:41:31.148236 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-tkg7w" Dec 03 06:41:31 crc kubenswrapper[4818]: I1203 06:41:31.350355 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-tkg7w"] Dec 03 06:41:31 crc kubenswrapper[4818]: W1203 06:41:31.354043 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf55ffc30_37c1_45e8_8e63_c447ca4d6e96.slice/crio-b26c5c88301e558159f4fdb784520993dc202fd050b45b7e99aa3c32fb21f9b8 WatchSource:0}: Error finding container b26c5c88301e558159f4fdb784520993dc202fd050b45b7e99aa3c32fb21f9b8: Status 404 returned error can't find the container with id b26c5c88301e558159f4fdb784520993dc202fd050b45b7e99aa3c32fb21f9b8 Dec 03 06:41:32 crc kubenswrapper[4818]: I1203 06:41:32.109603 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/860c17a3-a17f-4ac3-9310-c633df611f6d-memberlist\") pod \"speaker-4m9n7\" (UID: \"860c17a3-a17f-4ac3-9310-c633df611f6d\") " pod="metallb-system/speaker-4m9n7" Dec 03 06:41:32 crc kubenswrapper[4818]: I1203 06:41:32.122858 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/860c17a3-a17f-4ac3-9310-c633df611f6d-memberlist\") pod \"speaker-4m9n7\" (UID: \"860c17a3-a17f-4ac3-9310-c633df611f6d\") " pod="metallb-system/speaker-4m9n7" Dec 03 06:41:32 crc kubenswrapper[4818]: I1203 06:41:32.151527 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-7nr97" event={"ID":"bf628f10-0970-4651-8499-5302e78c8d6d","Type":"ContainerStarted","Data":"ae01cdf47b9a3a1f422ca30abe53ce643ace9a9d1220e40f23b8c9c267f62925"} Dec 03 06:41:32 crc kubenswrapper[4818]: I1203 06:41:32.151573 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-7nr97" event={"ID":"bf628f10-0970-4651-8499-5302e78c8d6d","Type":"ContainerStarted","Data":"e008edc4c32f4babc278e0f419035cf57ae673fcdc73ffaaeb79061812094242"} Dec 03 06:41:32 crc kubenswrapper[4818]: I1203 06:41:32.151614 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-7nr97" Dec 03 06:41:32 crc kubenswrapper[4818]: I1203 06:41:32.152790 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-tkg7w" event={"ID":"f55ffc30-37c1-45e8-8e63-c447ca4d6e96","Type":"ContainerStarted","Data":"b26c5c88301e558159f4fdb784520993dc202fd050b45b7e99aa3c32fb21f9b8"} Dec 03 06:41:32 crc kubenswrapper[4818]: I1203 06:41:32.161509 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-4m9n7" Dec 03 06:41:32 crc kubenswrapper[4818]: I1203 06:41:32.172522 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-7nr97" podStartSLOduration=2.172501513 podStartE2EDuration="2.172501513s" podCreationTimestamp="2025-12-03 06:41:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:41:32.17198656 +0000 UTC m=+849.863595312" watchObservedRunningTime="2025-12-03 06:41:32.172501513 +0000 UTC m=+849.864110275" Dec 03 06:41:32 crc kubenswrapper[4818]: W1203 06:41:32.194128 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod860c17a3_a17f_4ac3_9310_c633df611f6d.slice/crio-929694fd847e18914325c05b23efffe5e059c9831bccd2fe01f3b5d2e092d1bb WatchSource:0}: Error finding container 929694fd847e18914325c05b23efffe5e059c9831bccd2fe01f3b5d2e092d1bb: Status 404 returned error can't find the container with id 929694fd847e18914325c05b23efffe5e059c9831bccd2fe01f3b5d2e092d1bb Dec 03 06:41:33 crc kubenswrapper[4818]: I1203 06:41:33.158902 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4m9n7" event={"ID":"860c17a3-a17f-4ac3-9310-c633df611f6d","Type":"ContainerStarted","Data":"82b350bbe845d0d42af9416d387da443ccda109aa584de828086e698ea5fecc4"} Dec 03 06:41:33 crc kubenswrapper[4818]: I1203 06:41:33.158940 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4m9n7" event={"ID":"860c17a3-a17f-4ac3-9310-c633df611f6d","Type":"ContainerStarted","Data":"3f72039ea028a2e38ecf9876db9309f5230cd8599c73a5050a664eaa25aac0b4"} Dec 03 06:41:33 crc kubenswrapper[4818]: I1203 06:41:33.158953 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4m9n7" event={"ID":"860c17a3-a17f-4ac3-9310-c633df611f6d","Type":"ContainerStarted","Data":"929694fd847e18914325c05b23efffe5e059c9831bccd2fe01f3b5d2e092d1bb"} Dec 03 06:41:33 crc kubenswrapper[4818]: I1203 06:41:33.174695 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-4m9n7" podStartSLOduration=3.174678522 podStartE2EDuration="3.174678522s" podCreationTimestamp="2025-12-03 06:41:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:41:33.173534193 +0000 UTC m=+850.865142945" watchObservedRunningTime="2025-12-03 06:41:33.174678522 +0000 UTC m=+850.866287274" Dec 03 06:41:39 crc kubenswrapper[4818]: I1203 06:41:39.250352 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-tkg7w" event={"ID":"f55ffc30-37c1-45e8-8e63-c447ca4d6e96","Type":"ContainerStarted","Data":"6192d403f2c0298d10fc24e45f9bf73072ebe81829cf5fe6b23a6e202816ce1e"} Dec 03 06:41:39 crc kubenswrapper[4818]: I1203 06:41:39.252050 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-tkg7w" Dec 03 06:41:39 crc kubenswrapper[4818]: I1203 06:41:39.254852 4818 generic.go:334] "Generic (PLEG): container finished" podID="0a528d4d-3d84-40c2-a4b3-207001ba3579" containerID="fa98d0851cb3153eb041aefb1899c40778d6b005d31ed917be9aafad00f5c2c4" exitCode=0 Dec 03 06:41:39 crc kubenswrapper[4818]: I1203 06:41:39.254918 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hwmtn" event={"ID":"0a528d4d-3d84-40c2-a4b3-207001ba3579","Type":"ContainerDied","Data":"fa98d0851cb3153eb041aefb1899c40778d6b005d31ed917be9aafad00f5c2c4"} Dec 03 06:41:39 crc kubenswrapper[4818]: I1203 06:41:39.285328 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-tkg7w" podStartSLOduration=2.159584684 podStartE2EDuration="9.285302341s" podCreationTimestamp="2025-12-03 06:41:30 +0000 UTC" firstStartedPulling="2025-12-03 06:41:31.357070453 +0000 UTC m=+849.048679205" lastFinishedPulling="2025-12-03 06:41:38.4827881 +0000 UTC m=+856.174396862" observedRunningTime="2025-12-03 06:41:39.275500968 +0000 UTC m=+856.967109730" watchObservedRunningTime="2025-12-03 06:41:39.285302341 +0000 UTC m=+856.976911103" Dec 03 06:41:40 crc kubenswrapper[4818]: I1203 06:41:40.261921 4818 generic.go:334] "Generic (PLEG): container finished" podID="0a528d4d-3d84-40c2-a4b3-207001ba3579" containerID="14c767a5cb6428e6dc5c4ebffac49edf5a37c6aade30252c39c3d1d82d90ed28" exitCode=0 Dec 03 06:41:40 crc kubenswrapper[4818]: I1203 06:41:40.262130 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hwmtn" event={"ID":"0a528d4d-3d84-40c2-a4b3-207001ba3579","Type":"ContainerDied","Data":"14c767a5cb6428e6dc5c4ebffac49edf5a37c6aade30252c39c3d1d82d90ed28"} Dec 03 06:41:41 crc kubenswrapper[4818]: I1203 06:41:41.271437 4818 generic.go:334] "Generic (PLEG): container finished" podID="0a528d4d-3d84-40c2-a4b3-207001ba3579" containerID="0e46d456c675c79dfb7bb037baad6fb7c6c3f39aa9c2ac9857d7c4988c3ee30c" exitCode=0 Dec 03 06:41:41 crc kubenswrapper[4818]: I1203 06:41:41.272379 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hwmtn" event={"ID":"0a528d4d-3d84-40c2-a4b3-207001ba3579","Type":"ContainerDied","Data":"0e46d456c675c79dfb7bb037baad6fb7c6c3f39aa9c2ac9857d7c4988c3ee30c"} Dec 03 06:41:42 crc kubenswrapper[4818]: I1203 06:41:42.162499 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-4m9n7" Dec 03 06:41:42 crc kubenswrapper[4818]: I1203 06:41:42.167944 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-4m9n7" Dec 03 06:41:42 crc kubenswrapper[4818]: I1203 06:41:42.303746 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hwmtn" event={"ID":"0a528d4d-3d84-40c2-a4b3-207001ba3579","Type":"ContainerStarted","Data":"62d0267c5a110b505f3f644825befd358d648a7c91d9ae482cb70a01a1c00cf9"} Dec 03 06:41:42 crc kubenswrapper[4818]: I1203 06:41:42.303793 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hwmtn" event={"ID":"0a528d4d-3d84-40c2-a4b3-207001ba3579","Type":"ContainerStarted","Data":"e2e27de6ed67607d16bb9400e3f9cd9a5d751e46d33633cb45b67879fa121d10"} Dec 03 06:41:42 crc kubenswrapper[4818]: I1203 06:41:42.303809 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hwmtn" event={"ID":"0a528d4d-3d84-40c2-a4b3-207001ba3579","Type":"ContainerStarted","Data":"d7d6bb25e84608bb3fa60daf4ba645fae01b3bf1879729325c045338f41d1f38"} Dec 03 06:41:42 crc kubenswrapper[4818]: I1203 06:41:42.303840 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hwmtn" event={"ID":"0a528d4d-3d84-40c2-a4b3-207001ba3579","Type":"ContainerStarted","Data":"0829b59cd70ed7e29037622a66cc42416de45e612afd622867976db819140bb2"} Dec 03 06:41:42 crc kubenswrapper[4818]: I1203 06:41:42.303853 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hwmtn" event={"ID":"0a528d4d-3d84-40c2-a4b3-207001ba3579","Type":"ContainerStarted","Data":"597c871f02899614c2d512cf8374f5d51d1f2fab66b07a5cf5c75f165232f0d3"} Dec 03 06:41:43 crc kubenswrapper[4818]: I1203 06:41:43.313883 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hwmtn" event={"ID":"0a528d4d-3d84-40c2-a4b3-207001ba3579","Type":"ContainerStarted","Data":"438542d0251438fe1e4c28b5f0a5e65a16fe02fc4094aaec5fedd69e82770980"} Dec 03 06:41:43 crc kubenswrapper[4818]: I1203 06:41:43.314281 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:43 crc kubenswrapper[4818]: I1203 06:41:43.348727 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-hwmtn" podStartSLOduration=5.930833181 podStartE2EDuration="13.34870381s" podCreationTimestamp="2025-12-03 06:41:30 +0000 UTC" firstStartedPulling="2025-12-03 06:41:31.048102335 +0000 UTC m=+848.739711087" lastFinishedPulling="2025-12-03 06:41:38.465972954 +0000 UTC m=+856.157581716" observedRunningTime="2025-12-03 06:41:43.341751808 +0000 UTC m=+861.033360580" watchObservedRunningTime="2025-12-03 06:41:43.34870381 +0000 UTC m=+861.040312602" Dec 03 06:41:45 crc kubenswrapper[4818]: I1203 06:41:45.239556 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-jr8dk"] Dec 03 06:41:45 crc kubenswrapper[4818]: I1203 06:41:45.240787 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jr8dk" Dec 03 06:41:45 crc kubenswrapper[4818]: I1203 06:41:45.243142 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-gmkfx" Dec 03 06:41:45 crc kubenswrapper[4818]: I1203 06:41:45.243323 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 03 06:41:45 crc kubenswrapper[4818]: I1203 06:41:45.254626 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 03 06:41:45 crc kubenswrapper[4818]: I1203 06:41:45.265463 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-jr8dk"] Dec 03 06:41:45 crc kubenswrapper[4818]: I1203 06:41:45.351174 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqpd6\" (UniqueName: \"kubernetes.io/projected/f28e0eba-98e8-440c-8c04-91ce47ea304b-kube-api-access-qqpd6\") pod \"openstack-operator-index-jr8dk\" (UID: \"f28e0eba-98e8-440c-8c04-91ce47ea304b\") " pod="openstack-operators/openstack-operator-index-jr8dk" Dec 03 06:41:45 crc kubenswrapper[4818]: I1203 06:41:45.452918 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqpd6\" (UniqueName: \"kubernetes.io/projected/f28e0eba-98e8-440c-8c04-91ce47ea304b-kube-api-access-qqpd6\") pod \"openstack-operator-index-jr8dk\" (UID: \"f28e0eba-98e8-440c-8c04-91ce47ea304b\") " pod="openstack-operators/openstack-operator-index-jr8dk" Dec 03 06:41:45 crc kubenswrapper[4818]: I1203 06:41:45.487184 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqpd6\" (UniqueName: \"kubernetes.io/projected/f28e0eba-98e8-440c-8c04-91ce47ea304b-kube-api-access-qqpd6\") pod \"openstack-operator-index-jr8dk\" (UID: \"f28e0eba-98e8-440c-8c04-91ce47ea304b\") " pod="openstack-operators/openstack-operator-index-jr8dk" Dec 03 06:41:45 crc kubenswrapper[4818]: I1203 06:41:45.555633 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jr8dk" Dec 03 06:41:45 crc kubenswrapper[4818]: I1203 06:41:45.558234 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:45 crc kubenswrapper[4818]: I1203 06:41:45.623073 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:41:45 crc kubenswrapper[4818]: I1203 06:41:45.785046 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-jr8dk"] Dec 03 06:41:46 crc kubenswrapper[4818]: I1203 06:41:46.336100 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jr8dk" event={"ID":"f28e0eba-98e8-440c-8c04-91ce47ea304b","Type":"ContainerStarted","Data":"c0e55e702daaaf98e752fd745185d4eb25188166562fa50f86c4580380c9fbdf"} Dec 03 06:41:50 crc kubenswrapper[4818]: I1203 06:41:50.366410 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jr8dk" event={"ID":"f28e0eba-98e8-440c-8c04-91ce47ea304b","Type":"ContainerStarted","Data":"92793d441e000ce9eaa4a8cf097395e55f5f3cdbd79b6400faebfe7a4c683556"} Dec 03 06:41:50 crc kubenswrapper[4818]: I1203 06:41:50.392436 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-jr8dk" podStartSLOduration=1.846779218 podStartE2EDuration="5.392405855s" podCreationTimestamp="2025-12-03 06:41:45 +0000 UTC" firstStartedPulling="2025-12-03 06:41:45.789929804 +0000 UTC m=+863.481538556" lastFinishedPulling="2025-12-03 06:41:49.335556431 +0000 UTC m=+867.027165193" observedRunningTime="2025-12-03 06:41:50.385315609 +0000 UTC m=+868.076924391" watchObservedRunningTime="2025-12-03 06:41:50.392405855 +0000 UTC m=+868.084014637" Dec 03 06:41:50 crc kubenswrapper[4818]: I1203 06:41:50.714805 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-7nr97" Dec 03 06:41:51 crc kubenswrapper[4818]: I1203 06:41:51.158954 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-tkg7w" Dec 03 06:41:55 crc kubenswrapper[4818]: I1203 06:41:55.556351 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-jr8dk" Dec 03 06:41:55 crc kubenswrapper[4818]: I1203 06:41:55.557411 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-jr8dk" Dec 03 06:41:56 crc kubenswrapper[4818]: I1203 06:41:56.005007 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-jr8dk" Dec 03 06:41:56 crc kubenswrapper[4818]: I1203 06:41:56.440249 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-jr8dk" Dec 03 06:42:00 crc kubenswrapper[4818]: I1203 06:42:00.564932 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-hwmtn" Dec 03 06:42:03 crc kubenswrapper[4818]: I1203 06:42:03.873347 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk"] Dec 03 06:42:03 crc kubenswrapper[4818]: I1203 06:42:03.876186 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" Dec 03 06:42:03 crc kubenswrapper[4818]: I1203 06:42:03.883938 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-bwlmc" Dec 03 06:42:03 crc kubenswrapper[4818]: I1203 06:42:03.894238 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk"] Dec 03 06:42:04 crc kubenswrapper[4818]: I1203 06:42:04.042149 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw7vg\" (UniqueName: \"kubernetes.io/projected/413ec55a-90aa-4e14-a447-47595f913b9a-kube-api-access-pw7vg\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk\" (UID: \"413ec55a-90aa-4e14-a447-47595f913b9a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" Dec 03 06:42:04 crc kubenswrapper[4818]: I1203 06:42:04.042203 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/413ec55a-90aa-4e14-a447-47595f913b9a-bundle\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk\" (UID: \"413ec55a-90aa-4e14-a447-47595f913b9a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" Dec 03 06:42:04 crc kubenswrapper[4818]: I1203 06:42:04.042243 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/413ec55a-90aa-4e14-a447-47595f913b9a-util\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk\" (UID: \"413ec55a-90aa-4e14-a447-47595f913b9a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" Dec 03 06:42:04 crc kubenswrapper[4818]: I1203 06:42:04.143271 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/413ec55a-90aa-4e14-a447-47595f913b9a-util\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk\" (UID: \"413ec55a-90aa-4e14-a447-47595f913b9a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" Dec 03 06:42:04 crc kubenswrapper[4818]: I1203 06:42:04.143388 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw7vg\" (UniqueName: \"kubernetes.io/projected/413ec55a-90aa-4e14-a447-47595f913b9a-kube-api-access-pw7vg\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk\" (UID: \"413ec55a-90aa-4e14-a447-47595f913b9a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" Dec 03 06:42:04 crc kubenswrapper[4818]: I1203 06:42:04.143432 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/413ec55a-90aa-4e14-a447-47595f913b9a-bundle\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk\" (UID: \"413ec55a-90aa-4e14-a447-47595f913b9a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" Dec 03 06:42:04 crc kubenswrapper[4818]: I1203 06:42:04.143807 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/413ec55a-90aa-4e14-a447-47595f913b9a-util\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk\" (UID: \"413ec55a-90aa-4e14-a447-47595f913b9a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" Dec 03 06:42:04 crc kubenswrapper[4818]: I1203 06:42:04.143968 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/413ec55a-90aa-4e14-a447-47595f913b9a-bundle\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk\" (UID: \"413ec55a-90aa-4e14-a447-47595f913b9a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" Dec 03 06:42:04 crc kubenswrapper[4818]: I1203 06:42:04.165293 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw7vg\" (UniqueName: \"kubernetes.io/projected/413ec55a-90aa-4e14-a447-47595f913b9a-kube-api-access-pw7vg\") pod \"d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk\" (UID: \"413ec55a-90aa-4e14-a447-47595f913b9a\") " pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" Dec 03 06:42:04 crc kubenswrapper[4818]: I1203 06:42:04.217891 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" Dec 03 06:42:04 crc kubenswrapper[4818]: I1203 06:42:04.523377 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk"] Dec 03 06:42:04 crc kubenswrapper[4818]: W1203 06:42:04.531426 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod413ec55a_90aa_4e14_a447_47595f913b9a.slice/crio-abad56384e8146cdf4852ac5198d27f91463379b231e7ebf311a0a183eac954a WatchSource:0}: Error finding container abad56384e8146cdf4852ac5198d27f91463379b231e7ebf311a0a183eac954a: Status 404 returned error can't find the container with id abad56384e8146cdf4852ac5198d27f91463379b231e7ebf311a0a183eac954a Dec 03 06:42:05 crc kubenswrapper[4818]: I1203 06:42:05.472329 4818 generic.go:334] "Generic (PLEG): container finished" podID="413ec55a-90aa-4e14-a447-47595f913b9a" containerID="65e6e747493272a6f5ccbdb55ec0a44821344d8a0e46aa69af740e2a3e6fd852" exitCode=0 Dec 03 06:42:05 crc kubenswrapper[4818]: I1203 06:42:05.472381 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" event={"ID":"413ec55a-90aa-4e14-a447-47595f913b9a","Type":"ContainerDied","Data":"65e6e747493272a6f5ccbdb55ec0a44821344d8a0e46aa69af740e2a3e6fd852"} Dec 03 06:42:05 crc kubenswrapper[4818]: I1203 06:42:05.472412 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" event={"ID":"413ec55a-90aa-4e14-a447-47595f913b9a","Type":"ContainerStarted","Data":"abad56384e8146cdf4852ac5198d27f91463379b231e7ebf311a0a183eac954a"} Dec 03 06:42:06 crc kubenswrapper[4818]: I1203 06:42:06.482025 4818 generic.go:334] "Generic (PLEG): container finished" podID="413ec55a-90aa-4e14-a447-47595f913b9a" containerID="51753ec75f520c14f4802a51ce9289378cf5346589d5e8b022aa5fd5b55d3eec" exitCode=0 Dec 03 06:42:06 crc kubenswrapper[4818]: I1203 06:42:06.482085 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" event={"ID":"413ec55a-90aa-4e14-a447-47595f913b9a","Type":"ContainerDied","Data":"51753ec75f520c14f4802a51ce9289378cf5346589d5e8b022aa5fd5b55d3eec"} Dec 03 06:42:07 crc kubenswrapper[4818]: I1203 06:42:07.497141 4818 generic.go:334] "Generic (PLEG): container finished" podID="413ec55a-90aa-4e14-a447-47595f913b9a" containerID="1a9791a81cfc1895880a019319d443184a27e8203781f1cbc5de2a37e2c142bf" exitCode=0 Dec 03 06:42:07 crc kubenswrapper[4818]: I1203 06:42:07.497208 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" event={"ID":"413ec55a-90aa-4e14-a447-47595f913b9a","Type":"ContainerDied","Data":"1a9791a81cfc1895880a019319d443184a27e8203781f1cbc5de2a37e2c142bf"} Dec 03 06:42:08 crc kubenswrapper[4818]: I1203 06:42:08.783017 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" Dec 03 06:42:08 crc kubenswrapper[4818]: I1203 06:42:08.913304 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pw7vg\" (UniqueName: \"kubernetes.io/projected/413ec55a-90aa-4e14-a447-47595f913b9a-kube-api-access-pw7vg\") pod \"413ec55a-90aa-4e14-a447-47595f913b9a\" (UID: \"413ec55a-90aa-4e14-a447-47595f913b9a\") " Dec 03 06:42:08 crc kubenswrapper[4818]: I1203 06:42:08.913567 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/413ec55a-90aa-4e14-a447-47595f913b9a-bundle\") pod \"413ec55a-90aa-4e14-a447-47595f913b9a\" (UID: \"413ec55a-90aa-4e14-a447-47595f913b9a\") " Dec 03 06:42:08 crc kubenswrapper[4818]: I1203 06:42:08.913662 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/413ec55a-90aa-4e14-a447-47595f913b9a-util\") pod \"413ec55a-90aa-4e14-a447-47595f913b9a\" (UID: \"413ec55a-90aa-4e14-a447-47595f913b9a\") " Dec 03 06:42:08 crc kubenswrapper[4818]: I1203 06:42:08.914401 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/413ec55a-90aa-4e14-a447-47595f913b9a-bundle" (OuterVolumeSpecName: "bundle") pod "413ec55a-90aa-4e14-a447-47595f913b9a" (UID: "413ec55a-90aa-4e14-a447-47595f913b9a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:42:08 crc kubenswrapper[4818]: I1203 06:42:08.923027 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/413ec55a-90aa-4e14-a447-47595f913b9a-kube-api-access-pw7vg" (OuterVolumeSpecName: "kube-api-access-pw7vg") pod "413ec55a-90aa-4e14-a447-47595f913b9a" (UID: "413ec55a-90aa-4e14-a447-47595f913b9a"). InnerVolumeSpecName "kube-api-access-pw7vg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:42:08 crc kubenswrapper[4818]: I1203 06:42:08.944607 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/413ec55a-90aa-4e14-a447-47595f913b9a-util" (OuterVolumeSpecName: "util") pod "413ec55a-90aa-4e14-a447-47595f913b9a" (UID: "413ec55a-90aa-4e14-a447-47595f913b9a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:42:09 crc kubenswrapper[4818]: I1203 06:42:09.015520 4818 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/413ec55a-90aa-4e14-a447-47595f913b9a-util\") on node \"crc\" DevicePath \"\"" Dec 03 06:42:09 crc kubenswrapper[4818]: I1203 06:42:09.015598 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pw7vg\" (UniqueName: \"kubernetes.io/projected/413ec55a-90aa-4e14-a447-47595f913b9a-kube-api-access-pw7vg\") on node \"crc\" DevicePath \"\"" Dec 03 06:42:09 crc kubenswrapper[4818]: I1203 06:42:09.015623 4818 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/413ec55a-90aa-4e14-a447-47595f913b9a-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:42:09 crc kubenswrapper[4818]: I1203 06:42:09.517173 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" event={"ID":"413ec55a-90aa-4e14-a447-47595f913b9a","Type":"ContainerDied","Data":"abad56384e8146cdf4852ac5198d27f91463379b231e7ebf311a0a183eac954a"} Dec 03 06:42:09 crc kubenswrapper[4818]: I1203 06:42:09.517260 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abad56384e8146cdf4852ac5198d27f91463379b231e7ebf311a0a183eac954a" Dec 03 06:42:09 crc kubenswrapper[4818]: I1203 06:42:09.517308 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk" Dec 03 06:42:16 crc kubenswrapper[4818]: I1203 06:42:16.161509 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-555856bbb4-tlxbc"] Dec 03 06:42:16 crc kubenswrapper[4818]: E1203 06:42:16.163277 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="413ec55a-90aa-4e14-a447-47595f913b9a" containerName="pull" Dec 03 06:42:16 crc kubenswrapper[4818]: I1203 06:42:16.163324 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="413ec55a-90aa-4e14-a447-47595f913b9a" containerName="pull" Dec 03 06:42:16 crc kubenswrapper[4818]: E1203 06:42:16.163337 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="413ec55a-90aa-4e14-a447-47595f913b9a" containerName="extract" Dec 03 06:42:16 crc kubenswrapper[4818]: I1203 06:42:16.163347 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="413ec55a-90aa-4e14-a447-47595f913b9a" containerName="extract" Dec 03 06:42:16 crc kubenswrapper[4818]: E1203 06:42:16.163369 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="413ec55a-90aa-4e14-a447-47595f913b9a" containerName="util" Dec 03 06:42:16 crc kubenswrapper[4818]: I1203 06:42:16.163376 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="413ec55a-90aa-4e14-a447-47595f913b9a" containerName="util" Dec 03 06:42:16 crc kubenswrapper[4818]: I1203 06:42:16.163525 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="413ec55a-90aa-4e14-a447-47595f913b9a" containerName="extract" Dec 03 06:42:16 crc kubenswrapper[4818]: I1203 06:42:16.164095 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-555856bbb4-tlxbc" Dec 03 06:42:16 crc kubenswrapper[4818]: I1203 06:42:16.166323 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-fwf7f" Dec 03 06:42:16 crc kubenswrapper[4818]: I1203 06:42:16.185866 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-555856bbb4-tlxbc"] Dec 03 06:42:16 crc kubenswrapper[4818]: I1203 06:42:16.303862 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqls6\" (UniqueName: \"kubernetes.io/projected/69cb7eae-cb67-42ef-b36d-d6287f1f14d4-kube-api-access-fqls6\") pod \"openstack-operator-controller-operator-555856bbb4-tlxbc\" (UID: \"69cb7eae-cb67-42ef-b36d-d6287f1f14d4\") " pod="openstack-operators/openstack-operator-controller-operator-555856bbb4-tlxbc" Dec 03 06:42:16 crc kubenswrapper[4818]: I1203 06:42:16.405573 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqls6\" (UniqueName: \"kubernetes.io/projected/69cb7eae-cb67-42ef-b36d-d6287f1f14d4-kube-api-access-fqls6\") pod \"openstack-operator-controller-operator-555856bbb4-tlxbc\" (UID: \"69cb7eae-cb67-42ef-b36d-d6287f1f14d4\") " pod="openstack-operators/openstack-operator-controller-operator-555856bbb4-tlxbc" Dec 03 06:42:16 crc kubenswrapper[4818]: I1203 06:42:16.427677 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqls6\" (UniqueName: \"kubernetes.io/projected/69cb7eae-cb67-42ef-b36d-d6287f1f14d4-kube-api-access-fqls6\") pod \"openstack-operator-controller-operator-555856bbb4-tlxbc\" (UID: \"69cb7eae-cb67-42ef-b36d-d6287f1f14d4\") " pod="openstack-operators/openstack-operator-controller-operator-555856bbb4-tlxbc" Dec 03 06:42:16 crc kubenswrapper[4818]: I1203 06:42:16.487463 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-555856bbb4-tlxbc" Dec 03 06:42:16 crc kubenswrapper[4818]: I1203 06:42:16.734110 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-555856bbb4-tlxbc"] Dec 03 06:42:17 crc kubenswrapper[4818]: I1203 06:42:17.571360 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-555856bbb4-tlxbc" event={"ID":"69cb7eae-cb67-42ef-b36d-d6287f1f14d4","Type":"ContainerStarted","Data":"f16c99af8cdc407deab81c24c84dd31a84b8795f93daa368cab86678f26a7d56"} Dec 03 06:42:21 crc kubenswrapper[4818]: I1203 06:42:21.597989 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-555856bbb4-tlxbc" event={"ID":"69cb7eae-cb67-42ef-b36d-d6287f1f14d4","Type":"ContainerStarted","Data":"e2af432bc46506469fff22a01a7ba59109d285c848a8a106b08596818501c00b"} Dec 03 06:42:21 crc kubenswrapper[4818]: I1203 06:42:21.598635 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-555856bbb4-tlxbc" Dec 03 06:42:21 crc kubenswrapper[4818]: I1203 06:42:21.642173 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-555856bbb4-tlxbc" podStartSLOduration=1.514114813 podStartE2EDuration="5.642153334s" podCreationTimestamp="2025-12-03 06:42:16 +0000 UTC" firstStartedPulling="2025-12-03 06:42:16.745515083 +0000 UTC m=+894.437123825" lastFinishedPulling="2025-12-03 06:42:20.873553594 +0000 UTC m=+898.565162346" observedRunningTime="2025-12-03 06:42:21.636645097 +0000 UTC m=+899.328253889" watchObservedRunningTime="2025-12-03 06:42:21.642153334 +0000 UTC m=+899.333762096" Dec 03 06:42:25 crc kubenswrapper[4818]: I1203 06:42:25.303745 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-b4mm4"] Dec 03 06:42:25 crc kubenswrapper[4818]: I1203 06:42:25.305369 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b4mm4" Dec 03 06:42:25 crc kubenswrapper[4818]: I1203 06:42:25.314621 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b4mm4"] Dec 03 06:42:25 crc kubenswrapper[4818]: I1203 06:42:25.465512 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cff181b-16ba-4665-8111-264db353d63c-utilities\") pod \"community-operators-b4mm4\" (UID: \"4cff181b-16ba-4665-8111-264db353d63c\") " pod="openshift-marketplace/community-operators-b4mm4" Dec 03 06:42:25 crc kubenswrapper[4818]: I1203 06:42:25.465904 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h64c\" (UniqueName: \"kubernetes.io/projected/4cff181b-16ba-4665-8111-264db353d63c-kube-api-access-7h64c\") pod \"community-operators-b4mm4\" (UID: \"4cff181b-16ba-4665-8111-264db353d63c\") " pod="openshift-marketplace/community-operators-b4mm4" Dec 03 06:42:25 crc kubenswrapper[4818]: I1203 06:42:25.465938 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cff181b-16ba-4665-8111-264db353d63c-catalog-content\") pod \"community-operators-b4mm4\" (UID: \"4cff181b-16ba-4665-8111-264db353d63c\") " pod="openshift-marketplace/community-operators-b4mm4" Dec 03 06:42:25 crc kubenswrapper[4818]: I1203 06:42:25.566910 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cff181b-16ba-4665-8111-264db353d63c-utilities\") pod \"community-operators-b4mm4\" (UID: \"4cff181b-16ba-4665-8111-264db353d63c\") " pod="openshift-marketplace/community-operators-b4mm4" Dec 03 06:42:25 crc kubenswrapper[4818]: I1203 06:42:25.567228 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h64c\" (UniqueName: \"kubernetes.io/projected/4cff181b-16ba-4665-8111-264db353d63c-kube-api-access-7h64c\") pod \"community-operators-b4mm4\" (UID: \"4cff181b-16ba-4665-8111-264db353d63c\") " pod="openshift-marketplace/community-operators-b4mm4" Dec 03 06:42:25 crc kubenswrapper[4818]: I1203 06:42:25.567360 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cff181b-16ba-4665-8111-264db353d63c-catalog-content\") pod \"community-operators-b4mm4\" (UID: \"4cff181b-16ba-4665-8111-264db353d63c\") " pod="openshift-marketplace/community-operators-b4mm4" Dec 03 06:42:25 crc kubenswrapper[4818]: I1203 06:42:25.567444 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cff181b-16ba-4665-8111-264db353d63c-utilities\") pod \"community-operators-b4mm4\" (UID: \"4cff181b-16ba-4665-8111-264db353d63c\") " pod="openshift-marketplace/community-operators-b4mm4" Dec 03 06:42:25 crc kubenswrapper[4818]: I1203 06:42:25.567776 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cff181b-16ba-4665-8111-264db353d63c-catalog-content\") pod \"community-operators-b4mm4\" (UID: \"4cff181b-16ba-4665-8111-264db353d63c\") " pod="openshift-marketplace/community-operators-b4mm4" Dec 03 06:42:25 crc kubenswrapper[4818]: I1203 06:42:25.587408 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h64c\" (UniqueName: \"kubernetes.io/projected/4cff181b-16ba-4665-8111-264db353d63c-kube-api-access-7h64c\") pod \"community-operators-b4mm4\" (UID: \"4cff181b-16ba-4665-8111-264db353d63c\") " pod="openshift-marketplace/community-operators-b4mm4" Dec 03 06:42:25 crc kubenswrapper[4818]: I1203 06:42:25.623208 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b4mm4" Dec 03 06:42:25 crc kubenswrapper[4818]: I1203 06:42:25.903808 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b4mm4"] Dec 03 06:42:25 crc kubenswrapper[4818]: W1203 06:42:25.915548 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4cff181b_16ba_4665_8111_264db353d63c.slice/crio-631bb9510031d2990b2fd50520b0d41f50c56aca77de5f9daf2405c8bdd69172 WatchSource:0}: Error finding container 631bb9510031d2990b2fd50520b0d41f50c56aca77de5f9daf2405c8bdd69172: Status 404 returned error can't find the container with id 631bb9510031d2990b2fd50520b0d41f50c56aca77de5f9daf2405c8bdd69172 Dec 03 06:42:26 crc kubenswrapper[4818]: I1203 06:42:26.490629 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-555856bbb4-tlxbc" Dec 03 06:42:26 crc kubenswrapper[4818]: I1203 06:42:26.626920 4818 generic.go:334] "Generic (PLEG): container finished" podID="4cff181b-16ba-4665-8111-264db353d63c" containerID="b206e36ad0698ecf07ba24d2e196b0411d778c5178ef571a089b5d6aa4b4d3dd" exitCode=0 Dec 03 06:42:26 crc kubenswrapper[4818]: I1203 06:42:26.626959 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b4mm4" event={"ID":"4cff181b-16ba-4665-8111-264db353d63c","Type":"ContainerDied","Data":"b206e36ad0698ecf07ba24d2e196b0411d778c5178ef571a089b5d6aa4b4d3dd"} Dec 03 06:42:26 crc kubenswrapper[4818]: I1203 06:42:26.626985 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b4mm4" event={"ID":"4cff181b-16ba-4665-8111-264db353d63c","Type":"ContainerStarted","Data":"631bb9510031d2990b2fd50520b0d41f50c56aca77de5f9daf2405c8bdd69172"} Dec 03 06:42:27 crc kubenswrapper[4818]: I1203 06:42:27.635739 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b4mm4" event={"ID":"4cff181b-16ba-4665-8111-264db353d63c","Type":"ContainerStarted","Data":"5e2fd7dd2095192b975b3c23e4f6f55ca9033397aebfe3aac2a26ee17557f15b"} Dec 03 06:42:28 crc kubenswrapper[4818]: I1203 06:42:28.645260 4818 generic.go:334] "Generic (PLEG): container finished" podID="4cff181b-16ba-4665-8111-264db353d63c" containerID="5e2fd7dd2095192b975b3c23e4f6f55ca9033397aebfe3aac2a26ee17557f15b" exitCode=0 Dec 03 06:42:28 crc kubenswrapper[4818]: I1203 06:42:28.645361 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b4mm4" event={"ID":"4cff181b-16ba-4665-8111-264db353d63c","Type":"ContainerDied","Data":"5e2fd7dd2095192b975b3c23e4f6f55ca9033397aebfe3aac2a26ee17557f15b"} Dec 03 06:42:29 crc kubenswrapper[4818]: I1203 06:42:29.654749 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b4mm4" event={"ID":"4cff181b-16ba-4665-8111-264db353d63c","Type":"ContainerStarted","Data":"4b68e216cbb2f270d0e7cd3df174d340a26fe2824128428ecb2d26a8d6da9c27"} Dec 03 06:42:29 crc kubenswrapper[4818]: I1203 06:42:29.711340 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-b4mm4" podStartSLOduration=2.300866654 podStartE2EDuration="4.711325565s" podCreationTimestamp="2025-12-03 06:42:25 +0000 UTC" firstStartedPulling="2025-12-03 06:42:26.629199206 +0000 UTC m=+904.320807958" lastFinishedPulling="2025-12-03 06:42:29.039658107 +0000 UTC m=+906.731266869" observedRunningTime="2025-12-03 06:42:29.710448733 +0000 UTC m=+907.402057505" watchObservedRunningTime="2025-12-03 06:42:29.711325565 +0000 UTC m=+907.402934317" Dec 03 06:42:34 crc kubenswrapper[4818]: I1203 06:42:34.635886 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6vkcf"] Dec 03 06:42:34 crc kubenswrapper[4818]: I1203 06:42:34.637575 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6vkcf" Dec 03 06:42:34 crc kubenswrapper[4818]: I1203 06:42:34.664475 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6vkcf"] Dec 03 06:42:34 crc kubenswrapper[4818]: I1203 06:42:34.722232 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b02f78cd-9268-46b0-b66a-717793b48b5d-catalog-content\") pod \"redhat-marketplace-6vkcf\" (UID: \"b02f78cd-9268-46b0-b66a-717793b48b5d\") " pod="openshift-marketplace/redhat-marketplace-6vkcf" Dec 03 06:42:34 crc kubenswrapper[4818]: I1203 06:42:34.722307 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74thj\" (UniqueName: \"kubernetes.io/projected/b02f78cd-9268-46b0-b66a-717793b48b5d-kube-api-access-74thj\") pod \"redhat-marketplace-6vkcf\" (UID: \"b02f78cd-9268-46b0-b66a-717793b48b5d\") " pod="openshift-marketplace/redhat-marketplace-6vkcf" Dec 03 06:42:34 crc kubenswrapper[4818]: I1203 06:42:34.722340 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b02f78cd-9268-46b0-b66a-717793b48b5d-utilities\") pod \"redhat-marketplace-6vkcf\" (UID: \"b02f78cd-9268-46b0-b66a-717793b48b5d\") " pod="openshift-marketplace/redhat-marketplace-6vkcf" Dec 03 06:42:34 crc kubenswrapper[4818]: I1203 06:42:34.823757 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74thj\" (UniqueName: \"kubernetes.io/projected/b02f78cd-9268-46b0-b66a-717793b48b5d-kube-api-access-74thj\") pod \"redhat-marketplace-6vkcf\" (UID: \"b02f78cd-9268-46b0-b66a-717793b48b5d\") " pod="openshift-marketplace/redhat-marketplace-6vkcf" Dec 03 06:42:34 crc kubenswrapper[4818]: I1203 06:42:34.823829 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b02f78cd-9268-46b0-b66a-717793b48b5d-utilities\") pod \"redhat-marketplace-6vkcf\" (UID: \"b02f78cd-9268-46b0-b66a-717793b48b5d\") " pod="openshift-marketplace/redhat-marketplace-6vkcf" Dec 03 06:42:34 crc kubenswrapper[4818]: I1203 06:42:34.823882 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b02f78cd-9268-46b0-b66a-717793b48b5d-catalog-content\") pod \"redhat-marketplace-6vkcf\" (UID: \"b02f78cd-9268-46b0-b66a-717793b48b5d\") " pod="openshift-marketplace/redhat-marketplace-6vkcf" Dec 03 06:42:34 crc kubenswrapper[4818]: I1203 06:42:34.824320 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b02f78cd-9268-46b0-b66a-717793b48b5d-catalog-content\") pod \"redhat-marketplace-6vkcf\" (UID: \"b02f78cd-9268-46b0-b66a-717793b48b5d\") " pod="openshift-marketplace/redhat-marketplace-6vkcf" Dec 03 06:42:34 crc kubenswrapper[4818]: I1203 06:42:34.824388 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b02f78cd-9268-46b0-b66a-717793b48b5d-utilities\") pod \"redhat-marketplace-6vkcf\" (UID: \"b02f78cd-9268-46b0-b66a-717793b48b5d\") " pod="openshift-marketplace/redhat-marketplace-6vkcf" Dec 03 06:42:34 crc kubenswrapper[4818]: I1203 06:42:34.844055 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74thj\" (UniqueName: \"kubernetes.io/projected/b02f78cd-9268-46b0-b66a-717793b48b5d-kube-api-access-74thj\") pod \"redhat-marketplace-6vkcf\" (UID: \"b02f78cd-9268-46b0-b66a-717793b48b5d\") " pod="openshift-marketplace/redhat-marketplace-6vkcf" Dec 03 06:42:34 crc kubenswrapper[4818]: I1203 06:42:34.954059 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6vkcf" Dec 03 06:42:35 crc kubenswrapper[4818]: I1203 06:42:35.500809 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6vkcf"] Dec 03 06:42:35 crc kubenswrapper[4818]: I1203 06:42:35.623836 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-b4mm4" Dec 03 06:42:35 crc kubenswrapper[4818]: I1203 06:42:35.624834 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-b4mm4" Dec 03 06:42:35 crc kubenswrapper[4818]: I1203 06:42:35.681489 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-b4mm4" Dec 03 06:42:35 crc kubenswrapper[4818]: I1203 06:42:35.730095 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vkcf" event={"ID":"b02f78cd-9268-46b0-b66a-717793b48b5d","Type":"ContainerStarted","Data":"84535c69a5da35932867c01556d19a019d528feb9cb4f268fd7d5865c8810dd8"} Dec 03 06:42:35 crc kubenswrapper[4818]: I1203 06:42:35.780386 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-b4mm4" Dec 03 06:42:36 crc kubenswrapper[4818]: I1203 06:42:36.738267 4818 generic.go:334] "Generic (PLEG): container finished" podID="b02f78cd-9268-46b0-b66a-717793b48b5d" containerID="a97f75cf48c42883a2d240e78f3ed84eb6ab09960332d5bc5f75832d2a398586" exitCode=0 Dec 03 06:42:36 crc kubenswrapper[4818]: I1203 06:42:36.746020 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vkcf" event={"ID":"b02f78cd-9268-46b0-b66a-717793b48b5d","Type":"ContainerDied","Data":"a97f75cf48c42883a2d240e78f3ed84eb6ab09960332d5bc5f75832d2a398586"} Dec 03 06:42:38 crc kubenswrapper[4818]: I1203 06:42:38.041440 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b4mm4"] Dec 03 06:42:38 crc kubenswrapper[4818]: I1203 06:42:38.753960 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-b4mm4" podUID="4cff181b-16ba-4665-8111-264db353d63c" containerName="registry-server" containerID="cri-o://4b68e216cbb2f270d0e7cd3df174d340a26fe2824128428ecb2d26a8d6da9c27" gracePeriod=2 Dec 03 06:42:38 crc kubenswrapper[4818]: I1203 06:42:38.764713 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vkcf" event={"ID":"b02f78cd-9268-46b0-b66a-717793b48b5d","Type":"ContainerStarted","Data":"5d5943d870e8a32f386d26a15801a4e5e92ab3bb2e26f68513b0031e84f77454"} Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.156632 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b4mm4" Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.229899 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h64c\" (UniqueName: \"kubernetes.io/projected/4cff181b-16ba-4665-8111-264db353d63c-kube-api-access-7h64c\") pod \"4cff181b-16ba-4665-8111-264db353d63c\" (UID: \"4cff181b-16ba-4665-8111-264db353d63c\") " Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.230379 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cff181b-16ba-4665-8111-264db353d63c-utilities\") pod \"4cff181b-16ba-4665-8111-264db353d63c\" (UID: \"4cff181b-16ba-4665-8111-264db353d63c\") " Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.230460 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cff181b-16ba-4665-8111-264db353d63c-catalog-content\") pod \"4cff181b-16ba-4665-8111-264db353d63c\" (UID: \"4cff181b-16ba-4665-8111-264db353d63c\") " Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.231599 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cff181b-16ba-4665-8111-264db353d63c-utilities" (OuterVolumeSpecName: "utilities") pod "4cff181b-16ba-4665-8111-264db353d63c" (UID: "4cff181b-16ba-4665-8111-264db353d63c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.236958 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cff181b-16ba-4665-8111-264db353d63c-kube-api-access-7h64c" (OuterVolumeSpecName: "kube-api-access-7h64c") pod "4cff181b-16ba-4665-8111-264db353d63c" (UID: "4cff181b-16ba-4665-8111-264db353d63c"). InnerVolumeSpecName "kube-api-access-7h64c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.280934 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cff181b-16ba-4665-8111-264db353d63c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4cff181b-16ba-4665-8111-264db353d63c" (UID: "4cff181b-16ba-4665-8111-264db353d63c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.332133 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cff181b-16ba-4665-8111-264db353d63c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.332174 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cff181b-16ba-4665-8111-264db353d63c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.332189 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h64c\" (UniqueName: \"kubernetes.io/projected/4cff181b-16ba-4665-8111-264db353d63c-kube-api-access-7h64c\") on node \"crc\" DevicePath \"\"" Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.764579 4818 generic.go:334] "Generic (PLEG): container finished" podID="4cff181b-16ba-4665-8111-264db353d63c" containerID="4b68e216cbb2f270d0e7cd3df174d340a26fe2824128428ecb2d26a8d6da9c27" exitCode=0 Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.764625 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b4mm4" event={"ID":"4cff181b-16ba-4665-8111-264db353d63c","Type":"ContainerDied","Data":"4b68e216cbb2f270d0e7cd3df174d340a26fe2824128428ecb2d26a8d6da9c27"} Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.764677 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b4mm4" event={"ID":"4cff181b-16ba-4665-8111-264db353d63c","Type":"ContainerDied","Data":"631bb9510031d2990b2fd50520b0d41f50c56aca77de5f9daf2405c8bdd69172"} Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.764694 4818 scope.go:117] "RemoveContainer" containerID="4b68e216cbb2f270d0e7cd3df174d340a26fe2824128428ecb2d26a8d6da9c27" Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.764724 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b4mm4" Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.767016 4818 generic.go:334] "Generic (PLEG): container finished" podID="b02f78cd-9268-46b0-b66a-717793b48b5d" containerID="5d5943d870e8a32f386d26a15801a4e5e92ab3bb2e26f68513b0031e84f77454" exitCode=0 Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.767063 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vkcf" event={"ID":"b02f78cd-9268-46b0-b66a-717793b48b5d","Type":"ContainerDied","Data":"5d5943d870e8a32f386d26a15801a4e5e92ab3bb2e26f68513b0031e84f77454"} Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.782707 4818 scope.go:117] "RemoveContainer" containerID="5e2fd7dd2095192b975b3c23e4f6f55ca9033397aebfe3aac2a26ee17557f15b" Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.816428 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b4mm4"] Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.820088 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-b4mm4"] Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.828072 4818 scope.go:117] "RemoveContainer" containerID="b206e36ad0698ecf07ba24d2e196b0411d778c5178ef571a089b5d6aa4b4d3dd" Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.844298 4818 scope.go:117] "RemoveContainer" containerID="4b68e216cbb2f270d0e7cd3df174d340a26fe2824128428ecb2d26a8d6da9c27" Dec 03 06:42:39 crc kubenswrapper[4818]: E1203 06:42:39.844808 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b68e216cbb2f270d0e7cd3df174d340a26fe2824128428ecb2d26a8d6da9c27\": container with ID starting with 4b68e216cbb2f270d0e7cd3df174d340a26fe2824128428ecb2d26a8d6da9c27 not found: ID does not exist" containerID="4b68e216cbb2f270d0e7cd3df174d340a26fe2824128428ecb2d26a8d6da9c27" Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.844849 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b68e216cbb2f270d0e7cd3df174d340a26fe2824128428ecb2d26a8d6da9c27"} err="failed to get container status \"4b68e216cbb2f270d0e7cd3df174d340a26fe2824128428ecb2d26a8d6da9c27\": rpc error: code = NotFound desc = could not find container \"4b68e216cbb2f270d0e7cd3df174d340a26fe2824128428ecb2d26a8d6da9c27\": container with ID starting with 4b68e216cbb2f270d0e7cd3df174d340a26fe2824128428ecb2d26a8d6da9c27 not found: ID does not exist" Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.844869 4818 scope.go:117] "RemoveContainer" containerID="5e2fd7dd2095192b975b3c23e4f6f55ca9033397aebfe3aac2a26ee17557f15b" Dec 03 06:42:39 crc kubenswrapper[4818]: E1203 06:42:39.845170 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e2fd7dd2095192b975b3c23e4f6f55ca9033397aebfe3aac2a26ee17557f15b\": container with ID starting with 5e2fd7dd2095192b975b3c23e4f6f55ca9033397aebfe3aac2a26ee17557f15b not found: ID does not exist" containerID="5e2fd7dd2095192b975b3c23e4f6f55ca9033397aebfe3aac2a26ee17557f15b" Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.845188 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e2fd7dd2095192b975b3c23e4f6f55ca9033397aebfe3aac2a26ee17557f15b"} err="failed to get container status \"5e2fd7dd2095192b975b3c23e4f6f55ca9033397aebfe3aac2a26ee17557f15b\": rpc error: code = NotFound desc = could not find container \"5e2fd7dd2095192b975b3c23e4f6f55ca9033397aebfe3aac2a26ee17557f15b\": container with ID starting with 5e2fd7dd2095192b975b3c23e4f6f55ca9033397aebfe3aac2a26ee17557f15b not found: ID does not exist" Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.845199 4818 scope.go:117] "RemoveContainer" containerID="b206e36ad0698ecf07ba24d2e196b0411d778c5178ef571a089b5d6aa4b4d3dd" Dec 03 06:42:39 crc kubenswrapper[4818]: E1203 06:42:39.845420 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b206e36ad0698ecf07ba24d2e196b0411d778c5178ef571a089b5d6aa4b4d3dd\": container with ID starting with b206e36ad0698ecf07ba24d2e196b0411d778c5178ef571a089b5d6aa4b4d3dd not found: ID does not exist" containerID="b206e36ad0698ecf07ba24d2e196b0411d778c5178ef571a089b5d6aa4b4d3dd" Dec 03 06:42:39 crc kubenswrapper[4818]: I1203 06:42:39.845436 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b206e36ad0698ecf07ba24d2e196b0411d778c5178ef571a089b5d6aa4b4d3dd"} err="failed to get container status \"b206e36ad0698ecf07ba24d2e196b0411d778c5178ef571a089b5d6aa4b4d3dd\": rpc error: code = NotFound desc = could not find container \"b206e36ad0698ecf07ba24d2e196b0411d778c5178ef571a089b5d6aa4b4d3dd\": container with ID starting with b206e36ad0698ecf07ba24d2e196b0411d778c5178ef571a089b5d6aa4b4d3dd not found: ID does not exist" Dec 03 06:42:40 crc kubenswrapper[4818]: I1203 06:42:40.745972 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cff181b-16ba-4665-8111-264db353d63c" path="/var/lib/kubelet/pods/4cff181b-16ba-4665-8111-264db353d63c/volumes" Dec 03 06:42:40 crc kubenswrapper[4818]: I1203 06:42:40.776059 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vkcf" event={"ID":"b02f78cd-9268-46b0-b66a-717793b48b5d","Type":"ContainerStarted","Data":"ccfc503d381b73356e74e5dd6e58d14d99f8b895c607615c3330dbd414a9e0f0"} Dec 03 06:42:40 crc kubenswrapper[4818]: I1203 06:42:40.806942 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6vkcf" podStartSLOduration=3.275191379 podStartE2EDuration="6.806916224s" podCreationTimestamp="2025-12-03 06:42:34 +0000 UTC" firstStartedPulling="2025-12-03 06:42:36.74131906 +0000 UTC m=+914.432927812" lastFinishedPulling="2025-12-03 06:42:40.273043895 +0000 UTC m=+917.964652657" observedRunningTime="2025-12-03 06:42:40.802220717 +0000 UTC m=+918.493829469" watchObservedRunningTime="2025-12-03 06:42:40.806916224 +0000 UTC m=+918.498524976" Dec 03 06:42:44 crc kubenswrapper[4818]: I1203 06:42:44.955091 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6vkcf" Dec 03 06:42:44 crc kubenswrapper[4818]: I1203 06:42:44.956538 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6vkcf" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.010214 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6vkcf" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.256750 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-z6c4f"] Dec 03 06:42:45 crc kubenswrapper[4818]: E1203 06:42:45.256975 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cff181b-16ba-4665-8111-264db353d63c" containerName="extract-content" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.256986 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cff181b-16ba-4665-8111-264db353d63c" containerName="extract-content" Dec 03 06:42:45 crc kubenswrapper[4818]: E1203 06:42:45.256997 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cff181b-16ba-4665-8111-264db353d63c" containerName="registry-server" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.257003 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cff181b-16ba-4665-8111-264db353d63c" containerName="registry-server" Dec 03 06:42:45 crc kubenswrapper[4818]: E1203 06:42:45.257018 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cff181b-16ba-4665-8111-264db353d63c" containerName="extract-utilities" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.257024 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cff181b-16ba-4665-8111-264db353d63c" containerName="extract-utilities" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.257139 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cff181b-16ba-4665-8111-264db353d63c" containerName="registry-server" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.257692 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-z6c4f" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.260291 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-2sdt4" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.272702 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-z6c4f"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.279254 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-4hdr6"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.280440 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-4hdr6" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.284552 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-wndt9"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.284881 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-29v7h" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.285518 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wndt9" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.287291 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-4wbw4" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.309780 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-4hdr6"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.314212 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-2qvsd"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.315619 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-2qvsd" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.319992 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6hp9\" (UniqueName: \"kubernetes.io/projected/7c36b14e-4e1d-4956-a505-ecdffb9d9f9c-kube-api-access-f6hp9\") pod \"barbican-operator-controller-manager-7d9dfd778-z6c4f\" (UID: \"7c36b14e-4e1d-4956-a505-ecdffb9d9f9c\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-z6c4f" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.320112 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncrl8\" (UniqueName: \"kubernetes.io/projected/91e645ae-28ec-4d3b-8fe3-e75aa1acf5a5-kube-api-access-ncrl8\") pod \"designate-operator-controller-manager-78b4bc895b-wndt9\" (UID: \"91e645ae-28ec-4d3b-8fe3-e75aa1acf5a5\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wndt9" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.320218 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv8p5\" (UniqueName: \"kubernetes.io/projected/4130e7e0-4596-451f-b346-5afba08e9dea-kube-api-access-wv8p5\") pod \"cinder-operator-controller-manager-859b6ccc6-4hdr6\" (UID: \"4130e7e0-4596-451f-b346-5afba08e9dea\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-4hdr6" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.321487 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-pj55x" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.336333 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-wndt9"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.351496 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-2qvsd"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.369895 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gtmmz"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.371109 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gtmmz" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.372900 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-thzjt"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.373860 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-thzjt" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.374148 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-z6x2x" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.378971 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-8ggtx" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.392978 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gtmmz"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.413355 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-thzjt"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.419661 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.421130 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72nbw\" (UniqueName: \"kubernetes.io/projected/b29acffb-a7e4-4029-93fa-20f20de27dfe-kube-api-access-72nbw\") pod \"heat-operator-controller-manager-5f64f6f8bb-gtmmz\" (UID: \"b29acffb-a7e4-4029-93fa-20f20de27dfe\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gtmmz" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.421180 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6hp9\" (UniqueName: \"kubernetes.io/projected/7c36b14e-4e1d-4956-a505-ecdffb9d9f9c-kube-api-access-f6hp9\") pod \"barbican-operator-controller-manager-7d9dfd778-z6c4f\" (UID: \"7c36b14e-4e1d-4956-a505-ecdffb9d9f9c\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-z6c4f" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.421203 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncrl8\" (UniqueName: \"kubernetes.io/projected/91e645ae-28ec-4d3b-8fe3-e75aa1acf5a5-kube-api-access-ncrl8\") pod \"designate-operator-controller-manager-78b4bc895b-wndt9\" (UID: \"91e645ae-28ec-4d3b-8fe3-e75aa1acf5a5\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wndt9" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.421245 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-278sm\" (UniqueName: \"kubernetes.io/projected/6d5bf0d5-df89-4ab4-a711-e6f7f3d1123e-kube-api-access-278sm\") pod \"horizon-operator-controller-manager-68c6d99b8f-thzjt\" (UID: \"6d5bf0d5-df89-4ab4-a711-e6f7f3d1123e\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-thzjt" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.421272 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv8p5\" (UniqueName: \"kubernetes.io/projected/4130e7e0-4596-451f-b346-5afba08e9dea-kube-api-access-wv8p5\") pod \"cinder-operator-controller-manager-859b6ccc6-4hdr6\" (UID: \"4130e7e0-4596-451f-b346-5afba08e9dea\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-4hdr6" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.421298 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx6wv\" (UniqueName: \"kubernetes.io/projected/528c7c89-7980-4acc-9321-ff56c9d38489-kube-api-access-zx6wv\") pod \"glance-operator-controller-manager-77987cd8cd-2qvsd\" (UID: \"528c7c89-7980-4acc-9321-ff56c9d38489\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-2qvsd" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.421407 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.430489 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.430838 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-kmbtq" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.445927 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.457966 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6hp9\" (UniqueName: \"kubernetes.io/projected/7c36b14e-4e1d-4956-a505-ecdffb9d9f9c-kube-api-access-f6hp9\") pod \"barbican-operator-controller-manager-7d9dfd778-z6c4f\" (UID: \"7c36b14e-4e1d-4956-a505-ecdffb9d9f9c\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-z6c4f" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.463971 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv8p5\" (UniqueName: \"kubernetes.io/projected/4130e7e0-4596-451f-b346-5afba08e9dea-kube-api-access-wv8p5\") pod \"cinder-operator-controller-manager-859b6ccc6-4hdr6\" (UID: \"4130e7e0-4596-451f-b346-5afba08e9dea\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-4hdr6" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.475070 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncrl8\" (UniqueName: \"kubernetes.io/projected/91e645ae-28ec-4d3b-8fe3-e75aa1acf5a5-kube-api-access-ncrl8\") pod \"designate-operator-controller-manager-78b4bc895b-wndt9\" (UID: \"91e645ae-28ec-4d3b-8fe3-e75aa1acf5a5\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wndt9" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.488767 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-hpnhz"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.489807 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpnhz" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.493195 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-d8h5l" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.493385 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-hpnhz"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.506849 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-27nrr"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.509089 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-27nrr" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.513658 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-rwvc2" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.529531 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx6wv\" (UniqueName: \"kubernetes.io/projected/528c7c89-7980-4acc-9321-ff56c9d38489-kube-api-access-zx6wv\") pod \"glance-operator-controller-manager-77987cd8cd-2qvsd\" (UID: \"528c7c89-7980-4acc-9321-ff56c9d38489\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-2qvsd" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.529598 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72nbw\" (UniqueName: \"kubernetes.io/projected/b29acffb-a7e4-4029-93fa-20f20de27dfe-kube-api-access-72nbw\") pod \"heat-operator-controller-manager-5f64f6f8bb-gtmmz\" (UID: \"b29acffb-a7e4-4029-93fa-20f20de27dfe\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gtmmz" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.529626 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj878\" (UniqueName: \"kubernetes.io/projected/3d1565f0-62d4-4ecd-9d73-16338035196e-kube-api-access-zj878\") pod \"ironic-operator-controller-manager-6c548fd776-hpnhz\" (UID: \"3d1565f0-62d4-4ecd-9d73-16338035196e\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpnhz" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.529656 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47657\" (UniqueName: \"kubernetes.io/projected/a0b8fddb-4636-4f13-8d0b-3e3220721eca-kube-api-access-47657\") pod \"infra-operator-controller-manager-57548d458d-gmkzd\" (UID: \"a0b8fddb-4636-4f13-8d0b-3e3220721eca\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.529684 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert\") pod \"infra-operator-controller-manager-57548d458d-gmkzd\" (UID: \"a0b8fddb-4636-4f13-8d0b-3e3220721eca\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.529717 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-278sm\" (UniqueName: \"kubernetes.io/projected/6d5bf0d5-df89-4ab4-a711-e6f7f3d1123e-kube-api-access-278sm\") pod \"horizon-operator-controller-manager-68c6d99b8f-thzjt\" (UID: \"6d5bf0d5-df89-4ab4-a711-e6f7f3d1123e\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-thzjt" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.534706 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-s84mt"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.535641 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-s84mt" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.540700 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-7g9qm" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.549355 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-98z98"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.550450 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-98z98" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.552198 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-gmkn9" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.560984 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-27nrr"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.580113 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-z6c4f" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.580216 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-278sm\" (UniqueName: \"kubernetes.io/projected/6d5bf0d5-df89-4ab4-a711-e6f7f3d1123e-kube-api-access-278sm\") pod \"horizon-operator-controller-manager-68c6d99b8f-thzjt\" (UID: \"6d5bf0d5-df89-4ab4-a711-e6f7f3d1123e\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-thzjt" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.580574 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-98z98"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.586305 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72nbw\" (UniqueName: \"kubernetes.io/projected/b29acffb-a7e4-4029-93fa-20f20de27dfe-kube-api-access-72nbw\") pod \"heat-operator-controller-manager-5f64f6f8bb-gtmmz\" (UID: \"b29acffb-a7e4-4029-93fa-20f20de27dfe\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gtmmz" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.591259 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx6wv\" (UniqueName: \"kubernetes.io/projected/528c7c89-7980-4acc-9321-ff56c9d38489-kube-api-access-zx6wv\") pod \"glance-operator-controller-manager-77987cd8cd-2qvsd\" (UID: \"528c7c89-7980-4acc-9321-ff56c9d38489\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-2qvsd" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.593032 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-s84mt"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.604933 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b4thx"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.605885 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b4thx" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.611734 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-kgl5v" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.612127 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-4hdr6" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.615655 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b4thx"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.629903 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-xtq4s"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.631848 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert\") pod \"infra-operator-controller-manager-57548d458d-gmkzd\" (UID: \"a0b8fddb-4636-4f13-8d0b-3e3220721eca\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.631930 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcc48\" (UniqueName: \"kubernetes.io/projected/1bd3fb88-7598-4a28-90fe-324d855dc90e-kube-api-access-fcc48\") pod \"keystone-operator-controller-manager-7765d96ddf-27nrr\" (UID: \"1bd3fb88-7598-4a28-90fe-324d855dc90e\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-27nrr" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.631971 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgpfh\" (UniqueName: \"kubernetes.io/projected/df292cb6-9f13-4cec-a7e1-1ef3155a7b1f-kube-api-access-jgpfh\") pod \"mariadb-operator-controller-manager-56bbcc9d85-98z98\" (UID: \"df292cb6-9f13-4cec-a7e1-1ef3155a7b1f\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-98z98" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.632013 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgtjq\" (UniqueName: \"kubernetes.io/projected/0c360185-7fff-4054-a85e-ca2bbb1fa5e0-kube-api-access-hgtjq\") pod \"manila-operator-controller-manager-7c79b5df47-s84mt\" (UID: \"0c360185-7fff-4054-a85e-ca2bbb1fa5e0\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-s84mt" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.632069 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj878\" (UniqueName: \"kubernetes.io/projected/3d1565f0-62d4-4ecd-9d73-16338035196e-kube-api-access-zj878\") pod \"ironic-operator-controller-manager-6c548fd776-hpnhz\" (UID: \"3d1565f0-62d4-4ecd-9d73-16338035196e\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpnhz" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.632107 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47657\" (UniqueName: \"kubernetes.io/projected/a0b8fddb-4636-4f13-8d0b-3e3220721eca-kube-api-access-47657\") pod \"infra-operator-controller-manager-57548d458d-gmkzd\" (UID: \"a0b8fddb-4636-4f13-8d0b-3e3220721eca\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.632445 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xtq4s" Dec 03 06:42:45 crc kubenswrapper[4818]: E1203 06:42:45.632641 4818 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 06:42:45 crc kubenswrapper[4818]: E1203 06:42:45.632735 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert podName:a0b8fddb-4636-4f13-8d0b-3e3220721eca nodeName:}" failed. No retries permitted until 2025-12-03 06:42:46.132708027 +0000 UTC m=+923.824316779 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert") pod "infra-operator-controller-manager-57548d458d-gmkzd" (UID: "a0b8fddb-4636-4f13-8d0b-3e3220721eca") : secret "infra-operator-webhook-server-cert" not found Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.633353 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wndt9" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.652350 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-2qvsd" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.652486 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-gt2hd" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.665327 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-6rr64"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.699322 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-6rr64" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.721963 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-rm4f9" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.725617 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gtmmz" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.726995 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj878\" (UniqueName: \"kubernetes.io/projected/3d1565f0-62d4-4ecd-9d73-16338035196e-kube-api-access-zj878\") pod \"ironic-operator-controller-manager-6c548fd776-hpnhz\" (UID: \"3d1565f0-62d4-4ecd-9d73-16338035196e\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpnhz" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.737636 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-xtq4s"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.738782 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcc48\" (UniqueName: \"kubernetes.io/projected/1bd3fb88-7598-4a28-90fe-324d855dc90e-kube-api-access-fcc48\") pod \"keystone-operator-controller-manager-7765d96ddf-27nrr\" (UID: \"1bd3fb88-7598-4a28-90fe-324d855dc90e\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-27nrr" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.738873 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgpfh\" (UniqueName: \"kubernetes.io/projected/df292cb6-9f13-4cec-a7e1-1ef3155a7b1f-kube-api-access-jgpfh\") pod \"mariadb-operator-controller-manager-56bbcc9d85-98z98\" (UID: \"df292cb6-9f13-4cec-a7e1-1ef3155a7b1f\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-98z98" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.738912 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgtjq\" (UniqueName: \"kubernetes.io/projected/0c360185-7fff-4054-a85e-ca2bbb1fa5e0-kube-api-access-hgtjq\") pod \"manila-operator-controller-manager-7c79b5df47-s84mt\" (UID: \"0c360185-7fff-4054-a85e-ca2bbb1fa5e0\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-s84mt" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.739325 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-thzjt" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.741715 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47657\" (UniqueName: \"kubernetes.io/projected/a0b8fddb-4636-4f13-8d0b-3e3220721eca-kube-api-access-47657\") pod \"infra-operator-controller-manager-57548d458d-gmkzd\" (UID: \"a0b8fddb-4636-4f13-8d0b-3e3220721eca\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.801952 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-6rr64"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.808872 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.810005 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.812005 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-nchr7"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.813204 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nchr7" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.816133 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-df4sz" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.816263 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.816417 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-x7trc" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.840827 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvqj7\" (UniqueName: \"kubernetes.io/projected/1dc6c089-07b6-4818-a138-6b6a8c0cd1b5-kube-api-access-gvqj7\") pod \"nova-operator-controller-manager-697bc559fc-xtq4s\" (UID: \"1dc6c089-07b6-4818-a138-6b6a8c0cd1b5\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xtq4s" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.840945 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ght57\" (UniqueName: \"kubernetes.io/projected/45dcf421-464d-4eb9-a23f-9924ecf5bb03-kube-api-access-ght57\") pod \"octavia-operator-controller-manager-998648c74-6rr64\" (UID: \"45dcf421-464d-4eb9-a23f-9924ecf5bb03\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-6rr64" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.841005 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzcnn\" (UniqueName: \"kubernetes.io/projected/0d4b3be5-bb2e-4580-8142-60d15b16c36f-kube-api-access-nzcnn\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-b4thx\" (UID: \"0d4b3be5-bb2e-4580-8142-60d15b16c36f\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b4thx" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.841850 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpnhz" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.843139 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcc48\" (UniqueName: \"kubernetes.io/projected/1bd3fb88-7598-4a28-90fe-324d855dc90e-kube-api-access-fcc48\") pod \"keystone-operator-controller-manager-7765d96ddf-27nrr\" (UID: \"1bd3fb88-7598-4a28-90fe-324d855dc90e\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-27nrr" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.851778 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-nchr7"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.854800 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgtjq\" (UniqueName: \"kubernetes.io/projected/0c360185-7fff-4054-a85e-ca2bbb1fa5e0-kube-api-access-hgtjq\") pod \"manila-operator-controller-manager-7c79b5df47-s84mt\" (UID: \"0c360185-7fff-4054-a85e-ca2bbb1fa5e0\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-s84mt" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.858668 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgpfh\" (UniqueName: \"kubernetes.io/projected/df292cb6-9f13-4cec-a7e1-1ef3155a7b1f-kube-api-access-jgpfh\") pod \"mariadb-operator-controller-manager-56bbcc9d85-98z98\" (UID: \"df292cb6-9f13-4cec-a7e1-1ef3155a7b1f\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-98z98" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.859679 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-2d5xh"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.860654 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2d5xh" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.864267 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-27nrr" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.904070 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-ntml8" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.925934 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.934979 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-2d5xh"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.940956 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-r4b9n"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.942063 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-r4b9n" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.943456 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzcnn\" (UniqueName: \"kubernetes.io/projected/0d4b3be5-bb2e-4580-8142-60d15b16c36f-kube-api-access-nzcnn\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-b4thx\" (UID: \"0d4b3be5-bb2e-4580-8142-60d15b16c36f\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b4thx" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.943506 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvqj7\" (UniqueName: \"kubernetes.io/projected/1dc6c089-07b6-4818-a138-6b6a8c0cd1b5-kube-api-access-gvqj7\") pod \"nova-operator-controller-manager-697bc559fc-xtq4s\" (UID: \"1dc6c089-07b6-4818-a138-6b6a8c0cd1b5\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xtq4s" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.943561 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh87v\" (UniqueName: \"kubernetes.io/projected/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-kube-api-access-hh87v\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd445pgf\" (UID: \"f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.943589 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhr7t\" (UniqueName: \"kubernetes.io/projected/4b33f4e2-edb8-4e90-b73a-4dc2b13edd41-kube-api-access-nhr7t\") pod \"ovn-operator-controller-manager-b6456fdb6-nchr7\" (UID: \"4b33f4e2-edb8-4e90-b73a-4dc2b13edd41\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nchr7" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.943606 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd445pgf\" (UID: \"f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.943621 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ght57\" (UniqueName: \"kubernetes.io/projected/45dcf421-464d-4eb9-a23f-9924ecf5bb03-kube-api-access-ght57\") pod \"octavia-operator-controller-manager-998648c74-6rr64\" (UID: \"45dcf421-464d-4eb9-a23f-9924ecf5bb03\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-6rr64" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.957341 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-r4b9n"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.957678 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bjq7j"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.960837 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bjq7j" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.965652 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-n6clq" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.966399 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-s76lk" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.970559 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bjq7j"] Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.982336 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-s84mt" Dec 03 06:42:45 crc kubenswrapper[4818]: I1203 06:42:45.985357 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ght57\" (UniqueName: \"kubernetes.io/projected/45dcf421-464d-4eb9-a23f-9924ecf5bb03-kube-api-access-ght57\") pod \"octavia-operator-controller-manager-998648c74-6rr64\" (UID: \"45dcf421-464d-4eb9-a23f-9924ecf5bb03\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-6rr64" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:45.990337 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-98z98" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.042986 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-8hntr"] Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.044921 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8hntr" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.045562 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfk5n\" (UniqueName: \"kubernetes.io/projected/84dd0135-3ec6-4418-98f9-de8474662893-kube-api-access-pfk5n\") pod \"test-operator-controller-manager-5854674fcc-8hntr\" (UID: \"84dd0135-3ec6-4418-98f9-de8474662893\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-8hntr" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.045596 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlgnd\" (UniqueName: \"kubernetes.io/projected/e977170c-c8e9-4888-951b-bffa4d4d21d7-kube-api-access-mlgnd\") pod \"placement-operator-controller-manager-78f8948974-2d5xh\" (UID: \"e977170c-c8e9-4888-951b-bffa4d4d21d7\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-2d5xh" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.045642 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbbfn\" (UniqueName: \"kubernetes.io/projected/2e8a3376-fa5d-458d-a659-0c3cef8af10b-kube-api-access-nbbfn\") pod \"swift-operator-controller-manager-5f8c65bbfc-r4b9n\" (UID: \"2e8a3376-fa5d-458d-a659-0c3cef8af10b\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-r4b9n" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.045681 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh87v\" (UniqueName: \"kubernetes.io/projected/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-kube-api-access-hh87v\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd445pgf\" (UID: \"f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.045706 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhr7t\" (UniqueName: \"kubernetes.io/projected/4b33f4e2-edb8-4e90-b73a-4dc2b13edd41-kube-api-access-nhr7t\") pod \"ovn-operator-controller-manager-b6456fdb6-nchr7\" (UID: \"4b33f4e2-edb8-4e90-b73a-4dc2b13edd41\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nchr7" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.045728 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd445pgf\" (UID: \"f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.045765 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkl2p\" (UniqueName: \"kubernetes.io/projected/b78bb8f0-22f9-4b8e-93e2-23422db45354-kube-api-access-hkl2p\") pod \"telemetry-operator-controller-manager-76cc84c6bb-bjq7j\" (UID: \"b78bb8f0-22f9-4b8e-93e2-23422db45354\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bjq7j" Dec 03 06:42:46 crc kubenswrapper[4818]: E1203 06:42:46.047987 4818 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 06:42:46 crc kubenswrapper[4818]: E1203 06:42:46.048054 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert podName:f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34 nodeName:}" failed. No retries permitted until 2025-12-03 06:42:46.548036805 +0000 UTC m=+924.239645557 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" (UID: "f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.084444 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-4lkzs" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.110168 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvqj7\" (UniqueName: \"kubernetes.io/projected/1dc6c089-07b6-4818-a138-6b6a8c0cd1b5-kube-api-access-gvqj7\") pod \"nova-operator-controller-manager-697bc559fc-xtq4s\" (UID: \"1dc6c089-07b6-4818-a138-6b6a8c0cd1b5\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xtq4s" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.111068 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzcnn\" (UniqueName: \"kubernetes.io/projected/0d4b3be5-bb2e-4580-8142-60d15b16c36f-kube-api-access-nzcnn\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-b4thx\" (UID: \"0d4b3be5-bb2e-4580-8142-60d15b16c36f\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b4thx" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.111594 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-6rr64" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.120878 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-jd5hg"] Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.122185 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jd5hg" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.146889 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkl2p\" (UniqueName: \"kubernetes.io/projected/b78bb8f0-22f9-4b8e-93e2-23422db45354-kube-api-access-hkl2p\") pod \"telemetry-operator-controller-manager-76cc84c6bb-bjq7j\" (UID: \"b78bb8f0-22f9-4b8e-93e2-23422db45354\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bjq7j" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.147007 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfk5n\" (UniqueName: \"kubernetes.io/projected/84dd0135-3ec6-4418-98f9-de8474662893-kube-api-access-pfk5n\") pod \"test-operator-controller-manager-5854674fcc-8hntr\" (UID: \"84dd0135-3ec6-4418-98f9-de8474662893\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-8hntr" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.147036 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlgnd\" (UniqueName: \"kubernetes.io/projected/e977170c-c8e9-4888-951b-bffa4d4d21d7-kube-api-access-mlgnd\") pod \"placement-operator-controller-manager-78f8948974-2d5xh\" (UID: \"e977170c-c8e9-4888-951b-bffa4d4d21d7\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-2d5xh" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.147061 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert\") pod \"infra-operator-controller-manager-57548d458d-gmkzd\" (UID: \"a0b8fddb-4636-4f13-8d0b-3e3220721eca\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.147080 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbbfn\" (UniqueName: \"kubernetes.io/projected/2e8a3376-fa5d-458d-a659-0c3cef8af10b-kube-api-access-nbbfn\") pod \"swift-operator-controller-manager-5f8c65bbfc-r4b9n\" (UID: \"2e8a3376-fa5d-458d-a659-0c3cef8af10b\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-r4b9n" Dec 03 06:42:46 crc kubenswrapper[4818]: E1203 06:42:46.148418 4818 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 06:42:46 crc kubenswrapper[4818]: E1203 06:42:46.148468 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert podName:a0b8fddb-4636-4f13-8d0b-3e3220721eca nodeName:}" failed. No retries permitted until 2025-12-03 06:42:47.148453795 +0000 UTC m=+924.840062547 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert") pod "infra-operator-controller-manager-57548d458d-gmkzd" (UID: "a0b8fddb-4636-4f13-8d0b-3e3220721eca") : secret "infra-operator-webhook-server-cert" not found Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.153940 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-8hntr"] Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.158882 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-jmjwv" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.223413 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhr7t\" (UniqueName: \"kubernetes.io/projected/4b33f4e2-edb8-4e90-b73a-4dc2b13edd41-kube-api-access-nhr7t\") pod \"ovn-operator-controller-manager-b6456fdb6-nchr7\" (UID: \"4b33f4e2-edb8-4e90-b73a-4dc2b13edd41\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nchr7" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.225118 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh87v\" (UniqueName: \"kubernetes.io/projected/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-kube-api-access-hh87v\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd445pgf\" (UID: \"f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.317255 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srr8q\" (UniqueName: \"kubernetes.io/projected/56bb5f70-9dd8-433f-a974-0c95065c1dd1-kube-api-access-srr8q\") pod \"watcher-operator-controller-manager-769dc69bc-jd5hg\" (UID: \"56bb5f70-9dd8-433f-a974-0c95065c1dd1\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jd5hg" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.359420 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xtq4s" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.360286 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfk5n\" (UniqueName: \"kubernetes.io/projected/84dd0135-3ec6-4418-98f9-de8474662893-kube-api-access-pfk5n\") pod \"test-operator-controller-manager-5854674fcc-8hntr\" (UID: \"84dd0135-3ec6-4418-98f9-de8474662893\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-8hntr" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.361101 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkl2p\" (UniqueName: \"kubernetes.io/projected/b78bb8f0-22f9-4b8e-93e2-23422db45354-kube-api-access-hkl2p\") pod \"telemetry-operator-controller-manager-76cc84c6bb-bjq7j\" (UID: \"b78bb8f0-22f9-4b8e-93e2-23422db45354\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bjq7j" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.363392 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b4thx" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.419095 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6vkcf" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.421085 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srr8q\" (UniqueName: \"kubernetes.io/projected/56bb5f70-9dd8-433f-a974-0c95065c1dd1-kube-api-access-srr8q\") pod \"watcher-operator-controller-manager-769dc69bc-jd5hg\" (UID: \"56bb5f70-9dd8-433f-a974-0c95065c1dd1\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jd5hg" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.422382 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlgnd\" (UniqueName: \"kubernetes.io/projected/e977170c-c8e9-4888-951b-bffa4d4d21d7-kube-api-access-mlgnd\") pod \"placement-operator-controller-manager-78f8948974-2d5xh\" (UID: \"e977170c-c8e9-4888-951b-bffa4d4d21d7\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-2d5xh" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.434533 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbbfn\" (UniqueName: \"kubernetes.io/projected/2e8a3376-fa5d-458d-a659-0c3cef8af10b-kube-api-access-nbbfn\") pod \"swift-operator-controller-manager-5f8c65bbfc-r4b9n\" (UID: \"2e8a3376-fa5d-458d-a659-0c3cef8af10b\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-r4b9n" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.468186 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bjq7j" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.472924 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srr8q\" (UniqueName: \"kubernetes.io/projected/56bb5f70-9dd8-433f-a974-0c95065c1dd1-kube-api-access-srr8q\") pod \"watcher-operator-controller-manager-769dc69bc-jd5hg\" (UID: \"56bb5f70-9dd8-433f-a974-0c95065c1dd1\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jd5hg" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.494197 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-jd5hg"] Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.518351 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nchr7" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.592254 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6"] Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.607147 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.619925 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-xg6zs" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.620378 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.620669 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.623131 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2d5xh" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.628672 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8hntr" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.649457 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.650282 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsrvz\" (UniqueName: \"kubernetes.io/projected/3a635982-10c6-47cc-8462-74873b98410a-kube-api-access-nsrvz\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.652412 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd445pgf\" (UID: \"f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.652804 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:42:46 crc kubenswrapper[4818]: E1203 06:42:46.652892 4818 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 06:42:46 crc kubenswrapper[4818]: E1203 06:42:46.653177 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert podName:f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34 nodeName:}" failed. No retries permitted until 2025-12-03 06:42:47.652953106 +0000 UTC m=+925.344561858 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" (UID: "f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.660974 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6"] Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.701321 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-r4b9n" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.733977 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6vkcf"] Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.751788 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jd5hg" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.776109 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.776272 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.776656 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsrvz\" (UniqueName: \"kubernetes.io/projected/3a635982-10c6-47cc-8462-74873b98410a-kube-api-access-nsrvz\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:42:46 crc kubenswrapper[4818]: E1203 06:42:46.776838 4818 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 06:42:46 crc kubenswrapper[4818]: E1203 06:42:46.776912 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs podName:3a635982-10c6-47cc-8462-74873b98410a nodeName:}" failed. No retries permitted until 2025-12-03 06:42:47.276890779 +0000 UTC m=+924.968499531 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs") pod "openstack-operator-controller-manager-98d87bf44-8h8d6" (UID: "3a635982-10c6-47cc-8462-74873b98410a") : secret "webhook-server-cert" not found Dec 03 06:42:46 crc kubenswrapper[4818]: E1203 06:42:46.777142 4818 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 06:42:46 crc kubenswrapper[4818]: E1203 06:42:46.777218 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs podName:3a635982-10c6-47cc-8462-74873b98410a nodeName:}" failed. No retries permitted until 2025-12-03 06:42:47.277198936 +0000 UTC m=+924.968807688 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs") pod "openstack-operator-controller-manager-98d87bf44-8h8d6" (UID: "3a635982-10c6-47cc-8462-74873b98410a") : secret "metrics-server-cert" not found Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.882391 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsrvz\" (UniqueName: \"kubernetes.io/projected/3a635982-10c6-47cc-8462-74873b98410a-kube-api-access-nsrvz\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.896595 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q95v"] Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.897581 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q95v"] Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.897599 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-4hdr6"] Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.897706 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q95v" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.900444 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-g7pd7" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.926340 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-z6c4f"] Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.942541 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-wndt9"] Dec 03 06:42:46 crc kubenswrapper[4818]: W1203 06:42:46.950953 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4130e7e0_4596_451f_b346_5afba08e9dea.slice/crio-34fe0427ea142de0e784c3835dade967911edf19834bae882716a594125e56fd WatchSource:0}: Error finding container 34fe0427ea142de0e784c3835dade967911edf19834bae882716a594125e56fd: Status 404 returned error can't find the container with id 34fe0427ea142de0e784c3835dade967911edf19834bae882716a594125e56fd Dec 03 06:42:46 crc kubenswrapper[4818]: W1203 06:42:46.976025 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c36b14e_4e1d_4956_a505_ecdffb9d9f9c.slice/crio-2a9470e90fb35faddff02787443f9a0c4e41507b9928dd36c76fac2651cd3af9 WatchSource:0}: Error finding container 2a9470e90fb35faddff02787443f9a0c4e41507b9928dd36c76fac2651cd3af9: Status 404 returned error can't find the container with id 2a9470e90fb35faddff02787443f9a0c4e41507b9928dd36c76fac2651cd3af9 Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.989392 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89md4\" (UniqueName: \"kubernetes.io/projected/5a1859a2-fa2e-4d8c-bd64-7b63297f258c-kube-api-access-89md4\") pod \"rabbitmq-cluster-operator-manager-668c99d594-5q95v\" (UID: \"5a1859a2-fa2e-4d8c-bd64-7b63297f258c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q95v" Dec 03 06:42:46 crc kubenswrapper[4818]: I1203 06:42:46.998094 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-4hdr6" event={"ID":"4130e7e0-4596-451f-b346-5afba08e9dea","Type":"ContainerStarted","Data":"34fe0427ea142de0e784c3835dade967911edf19834bae882716a594125e56fd"} Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.094567 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89md4\" (UniqueName: \"kubernetes.io/projected/5a1859a2-fa2e-4d8c-bd64-7b63297f258c-kube-api-access-89md4\") pod \"rabbitmq-cluster-operator-manager-668c99d594-5q95v\" (UID: \"5a1859a2-fa2e-4d8c-bd64-7b63297f258c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q95v" Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.125708 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89md4\" (UniqueName: \"kubernetes.io/projected/5a1859a2-fa2e-4d8c-bd64-7b63297f258c-kube-api-access-89md4\") pod \"rabbitmq-cluster-operator-manager-668c99d594-5q95v\" (UID: \"5a1859a2-fa2e-4d8c-bd64-7b63297f258c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q95v" Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.211476 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert\") pod \"infra-operator-controller-manager-57548d458d-gmkzd\" (UID: \"a0b8fddb-4636-4f13-8d0b-3e3220721eca\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" Dec 03 06:42:47 crc kubenswrapper[4818]: E1203 06:42:47.211734 4818 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 06:42:47 crc kubenswrapper[4818]: E1203 06:42:47.211787 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert podName:a0b8fddb-4636-4f13-8d0b-3e3220721eca nodeName:}" failed. No retries permitted until 2025-12-03 06:42:49.211771942 +0000 UTC m=+926.903380694 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert") pod "infra-operator-controller-manager-57548d458d-gmkzd" (UID: "a0b8fddb-4636-4f13-8d0b-3e3220721eca") : secret "infra-operator-webhook-server-cert" not found Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.282090 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q95v" Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.331505 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.331585 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:42:47 crc kubenswrapper[4818]: E1203 06:42:47.331693 4818 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 06:42:47 crc kubenswrapper[4818]: E1203 06:42:47.331713 4818 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 06:42:47 crc kubenswrapper[4818]: E1203 06:42:47.331763 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs podName:3a635982-10c6-47cc-8462-74873b98410a nodeName:}" failed. No retries permitted until 2025-12-03 06:42:48.331745677 +0000 UTC m=+926.023354419 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs") pod "openstack-operator-controller-manager-98d87bf44-8h8d6" (UID: "3a635982-10c6-47cc-8462-74873b98410a") : secret "webhook-server-cert" not found Dec 03 06:42:47 crc kubenswrapper[4818]: E1203 06:42:47.331779 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs podName:3a635982-10c6-47cc-8462-74873b98410a nodeName:}" failed. No retries permitted until 2025-12-03 06:42:48.331773248 +0000 UTC m=+926.023382000 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs") pod "openstack-operator-controller-manager-98d87bf44-8h8d6" (UID: "3a635982-10c6-47cc-8462-74873b98410a") : secret "metrics-server-cert" not found Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.451596 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gtmmz"] Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.455948 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-27nrr"] Dec 03 06:42:47 crc kubenswrapper[4818]: W1203 06:42:47.463568 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb29acffb_a7e4_4029_93fa_20f20de27dfe.slice/crio-8cff2a3ad17042e0c3ec1698ff045b3cd03fe7dae533f603bb72fd5ad026b1d4 WatchSource:0}: Error finding container 8cff2a3ad17042e0c3ec1698ff045b3cd03fe7dae533f603bb72fd5ad026b1d4: Status 404 returned error can't find the container with id 8cff2a3ad17042e0c3ec1698ff045b3cd03fe7dae533f603bb72fd5ad026b1d4 Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.471675 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-hpnhz"] Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.517013 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-thzjt"] Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.705893 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b4thx"] Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.722186 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-xtq4s"] Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.727936 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-98z98"] Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.732607 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-s84mt"] Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.739745 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-2qvsd"] Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.748008 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-6rr64"] Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.748619 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd445pgf\" (UID: \"f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" Dec 03 06:42:47 crc kubenswrapper[4818]: E1203 06:42:47.748753 4818 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 06:42:47 crc kubenswrapper[4818]: E1203 06:42:47.748796 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert podName:f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34 nodeName:}" failed. No retries permitted until 2025-12-03 06:42:49.748781688 +0000 UTC m=+927.440390440 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" (UID: "f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.885594 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-r4b9n"] Dec 03 06:42:47 crc kubenswrapper[4818]: W1203 06:42:47.896408 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e8a3376_fa5d_458d_a659_0c3cef8af10b.slice/crio-5ed244709a98cde35ed4519acb938ea2d1f0ef3fa18694a31a02639f0d7e5392 WatchSource:0}: Error finding container 5ed244709a98cde35ed4519acb938ea2d1f0ef3fa18694a31a02639f0d7e5392: Status 404 returned error can't find the container with id 5ed244709a98cde35ed4519acb938ea2d1f0ef3fa18694a31a02639f0d7e5392 Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.904177 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bjq7j"] Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.909084 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-8hntr"] Dec 03 06:42:47 crc kubenswrapper[4818]: W1203 06:42:47.913901 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84dd0135_3ec6_4418_98f9_de8474662893.slice/crio-49858f8c8fe9258e35c19f0ca027e3b6f4ff3beba6f7589e1329bd55c2314307 WatchSource:0}: Error finding container 49858f8c8fe9258e35c19f0ca027e3b6f4ff3beba6f7589e1329bd55c2314307: Status 404 returned error can't find the container with id 49858f8c8fe9258e35c19f0ca027e3b6f4ff3beba6f7589e1329bd55c2314307 Dec 03 06:42:47 crc kubenswrapper[4818]: I1203 06:42:47.917224 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-jd5hg"] Dec 03 06:42:47 crc kubenswrapper[4818]: E1203 06:42:47.917266 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pfk5n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-8hntr_openstack-operators(84dd0135-3ec6-4418-98f9-de8474662893): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 06:42:47 crc kubenswrapper[4818]: E1203 06:42:47.917686 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-srr8q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-jd5hg_openstack-operators(56bb5f70-9dd8-433f-a974-0c95065c1dd1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 06:42:47 crc kubenswrapper[4818]: E1203 06:42:47.919194 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-srr8q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-jd5hg_openstack-operators(56bb5f70-9dd8-433f-a974-0c95065c1dd1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 06:42:47 crc kubenswrapper[4818]: E1203 06:42:47.919288 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hkl2p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-bjq7j_openstack-operators(b78bb8f0-22f9-4b8e-93e2-23422db45354): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 06:42:47 crc kubenswrapper[4818]: E1203 06:42:47.919589 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pfk5n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-8hntr_openstack-operators(84dd0135-3ec6-4418-98f9-de8474662893): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 06:42:47 crc kubenswrapper[4818]: E1203 06:42:47.920968 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jd5hg" podUID="56bb5f70-9dd8-433f-a974-0c95065c1dd1" Dec 03 06:42:47 crc kubenswrapper[4818]: E1203 06:42:47.920949 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8hntr" podUID="84dd0135-3ec6-4418-98f9-de8474662893" Dec 03 06:42:47 crc kubenswrapper[4818]: E1203 06:42:47.921469 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hkl2p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-bjq7j_openstack-operators(b78bb8f0-22f9-4b8e-93e2-23422db45354): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 06:42:47 crc kubenswrapper[4818]: E1203 06:42:47.922603 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bjq7j" podUID="b78bb8f0-22f9-4b8e-93e2-23422db45354" Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.005408 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-thzjt" event={"ID":"6d5bf0d5-df89-4ab4-a711-e6f7f3d1123e","Type":"ContainerStarted","Data":"ae2d6b70c06385891b6bc5633777c472f9b0d533adbb9e808e60caacb5069942"} Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.009834 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b4thx" event={"ID":"0d4b3be5-bb2e-4580-8142-60d15b16c36f","Type":"ContainerStarted","Data":"77b84173554cadbceb1ea483453d2fc6cf6997631ea3f5ddd1b816a8cec06f32"} Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.011038 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jd5hg" event={"ID":"56bb5f70-9dd8-433f-a974-0c95065c1dd1","Type":"ContainerStarted","Data":"68886300b98c71576da4ea0aadb4abe1a0dc8810889bb72bd346f4f151ce4656"} Dec 03 06:42:48 crc kubenswrapper[4818]: E1203 06:42:48.012843 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jd5hg" podUID="56bb5f70-9dd8-433f-a974-0c95065c1dd1" Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.012972 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-z6c4f" event={"ID":"7c36b14e-4e1d-4956-a505-ecdffb9d9f9c","Type":"ContainerStarted","Data":"2a9470e90fb35faddff02787443f9a0c4e41507b9928dd36c76fac2651cd3af9"} Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.013650 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xtq4s" event={"ID":"1dc6c089-07b6-4818-a138-6b6a8c0cd1b5","Type":"ContainerStarted","Data":"59dbee3ab1626dde171235f2b31d574007665903b12e95bb81be1ec612dc1318"} Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.015401 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-2qvsd" event={"ID":"528c7c89-7980-4acc-9321-ff56c9d38489","Type":"ContainerStarted","Data":"d1e5bdc0e2abb2251643bfdce08298e58833fde18edb467499865a881f7b3110"} Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.016336 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-nchr7"] Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.017628 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-27nrr" event={"ID":"1bd3fb88-7598-4a28-90fe-324d855dc90e","Type":"ContainerStarted","Data":"9cdcfc7059507dea2de51388559ce1a183e5244d70857b5423424613429b50b2"} Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.019753 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gtmmz" event={"ID":"b29acffb-a7e4-4029-93fa-20f20de27dfe","Type":"ContainerStarted","Data":"8cff2a3ad17042e0c3ec1698ff045b3cd03fe7dae533f603bb72fd5ad026b1d4"} Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.020651 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-r4b9n" event={"ID":"2e8a3376-fa5d-458d-a659-0c3cef8af10b","Type":"ContainerStarted","Data":"5ed244709a98cde35ed4519acb938ea2d1f0ef3fa18694a31a02639f0d7e5392"} Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.021321 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-2d5xh"] Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.022719 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-6rr64" event={"ID":"45dcf421-464d-4eb9-a23f-9924ecf5bb03","Type":"ContainerStarted","Data":"59f67f3cd94ea034a0870bca3737155cfe37184ad849a669b077b301a8a73e0c"} Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.029250 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8hntr" event={"ID":"84dd0135-3ec6-4418-98f9-de8474662893","Type":"ContainerStarted","Data":"49858f8c8fe9258e35c19f0ca027e3b6f4ff3beba6f7589e1329bd55c2314307"} Dec 03 06:42:48 crc kubenswrapper[4818]: E1203 06:42:48.032663 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nhr7t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-nchr7_openstack-operators(4b33f4e2-edb8-4e90-b73a-4dc2b13edd41): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 06:42:48 crc kubenswrapper[4818]: E1203 06:42:48.033131 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8hntr" podUID="84dd0135-3ec6-4418-98f9-de8474662893" Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.033909 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wndt9" event={"ID":"91e645ae-28ec-4d3b-8fe3-e75aa1acf5a5","Type":"ContainerStarted","Data":"7119bb43573dbb67b380522213561078552e0efc7eaa1ccc3ce3b22e706f9507"} Dec 03 06:42:48 crc kubenswrapper[4818]: E1203 06:42:48.034623 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nhr7t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-nchr7_openstack-operators(4b33f4e2-edb8-4e90-b73a-4dc2b13edd41): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.035099 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bjq7j" event={"ID":"b78bb8f0-22f9-4b8e-93e2-23422db45354","Type":"ContainerStarted","Data":"d4aa6b333095014c0de7a7cd7040ec7fcf17b4de9171e040980658cadf376670"} Dec 03 06:42:48 crc kubenswrapper[4818]: E1203 06:42:48.035862 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nchr7" podUID="4b33f4e2-edb8-4e90-b73a-4dc2b13edd41" Dec 03 06:42:48 crc kubenswrapper[4818]: E1203 06:42:48.037549 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bjq7j" podUID="b78bb8f0-22f9-4b8e-93e2-23422db45354" Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.037934 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-s84mt" event={"ID":"0c360185-7fff-4054-a85e-ca2bbb1fa5e0","Type":"ContainerStarted","Data":"9022425640c21f734c46f2c01b96c4176479570f8acbb07ae1504ca27f10ce69"} Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.039679 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpnhz" event={"ID":"3d1565f0-62d4-4ecd-9d73-16338035196e","Type":"ContainerStarted","Data":"1a134de4ef63c41f9dc6bc004e40ae217caa069c11c535e03a15657920dff92d"} Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.044662 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6vkcf" podUID="b02f78cd-9268-46b0-b66a-717793b48b5d" containerName="registry-server" containerID="cri-o://ccfc503d381b73356e74e5dd6e58d14d99f8b895c607615c3330dbd414a9e0f0" gracePeriod=2 Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.044755 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-98z98" event={"ID":"df292cb6-9f13-4cec-a7e1-1ef3155a7b1f","Type":"ContainerStarted","Data":"8bfb6afc011f5bb049160b89bea1a615df1cdd69e363a2ad40936796e46c7215"} Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.051652 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q95v"] Dec 03 06:42:48 crc kubenswrapper[4818]: E1203 06:42:48.067773 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-89md4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-5q95v_openstack-operators(5a1859a2-fa2e-4d8c-bd64-7b63297f258c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 06:42:48 crc kubenswrapper[4818]: E1203 06:42:48.069887 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q95v" podUID="5a1859a2-fa2e-4d8c-bd64-7b63297f258c" Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.359352 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:42:48 crc kubenswrapper[4818]: E1203 06:42:48.359519 4818 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.359550 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:42:48 crc kubenswrapper[4818]: E1203 06:42:48.359584 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs podName:3a635982-10c6-47cc-8462-74873b98410a nodeName:}" failed. No retries permitted until 2025-12-03 06:42:50.359565304 +0000 UTC m=+928.051174056 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs") pod "openstack-operator-controller-manager-98d87bf44-8h8d6" (UID: "3a635982-10c6-47cc-8462-74873b98410a") : secret "metrics-server-cert" not found Dec 03 06:42:48 crc kubenswrapper[4818]: E1203 06:42:48.359770 4818 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 06:42:48 crc kubenswrapper[4818]: E1203 06:42:48.359882 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs podName:3a635982-10c6-47cc-8462-74873b98410a nodeName:}" failed. No retries permitted until 2025-12-03 06:42:50.359858971 +0000 UTC m=+928.051467733 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs") pod "openstack-operator-controller-manager-98d87bf44-8h8d6" (UID: "3a635982-10c6-47cc-8462-74873b98410a") : secret "webhook-server-cert" not found Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.721765 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6vkcf" Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.869028 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b02f78cd-9268-46b0-b66a-717793b48b5d-utilities\") pod \"b02f78cd-9268-46b0-b66a-717793b48b5d\" (UID: \"b02f78cd-9268-46b0-b66a-717793b48b5d\") " Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.869391 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b02f78cd-9268-46b0-b66a-717793b48b5d-catalog-content\") pod \"b02f78cd-9268-46b0-b66a-717793b48b5d\" (UID: \"b02f78cd-9268-46b0-b66a-717793b48b5d\") " Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.869530 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74thj\" (UniqueName: \"kubernetes.io/projected/b02f78cd-9268-46b0-b66a-717793b48b5d-kube-api-access-74thj\") pod \"b02f78cd-9268-46b0-b66a-717793b48b5d\" (UID: \"b02f78cd-9268-46b0-b66a-717793b48b5d\") " Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.898129 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b02f78cd-9268-46b0-b66a-717793b48b5d-utilities" (OuterVolumeSpecName: "utilities") pod "b02f78cd-9268-46b0-b66a-717793b48b5d" (UID: "b02f78cd-9268-46b0-b66a-717793b48b5d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.905084 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b02f78cd-9268-46b0-b66a-717793b48b5d-kube-api-access-74thj" (OuterVolumeSpecName: "kube-api-access-74thj") pod "b02f78cd-9268-46b0-b66a-717793b48b5d" (UID: "b02f78cd-9268-46b0-b66a-717793b48b5d"). InnerVolumeSpecName "kube-api-access-74thj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.911139 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b02f78cd-9268-46b0-b66a-717793b48b5d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b02f78cd-9268-46b0-b66a-717793b48b5d" (UID: "b02f78cd-9268-46b0-b66a-717793b48b5d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.972020 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74thj\" (UniqueName: \"kubernetes.io/projected/b02f78cd-9268-46b0-b66a-717793b48b5d-kube-api-access-74thj\") on node \"crc\" DevicePath \"\"" Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.972053 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b02f78cd-9268-46b0-b66a-717793b48b5d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:42:48 crc kubenswrapper[4818]: I1203 06:42:48.972062 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b02f78cd-9268-46b0-b66a-717793b48b5d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.059068 4818 generic.go:334] "Generic (PLEG): container finished" podID="b02f78cd-9268-46b0-b66a-717793b48b5d" containerID="ccfc503d381b73356e74e5dd6e58d14d99f8b895c607615c3330dbd414a9e0f0" exitCode=0 Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.059190 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vkcf" event={"ID":"b02f78cd-9268-46b0-b66a-717793b48b5d","Type":"ContainerDied","Data":"ccfc503d381b73356e74e5dd6e58d14d99f8b895c607615c3330dbd414a9e0f0"} Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.059241 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vkcf" event={"ID":"b02f78cd-9268-46b0-b66a-717793b48b5d","Type":"ContainerDied","Data":"84535c69a5da35932867c01556d19a019d528feb9cb4f268fd7d5865c8810dd8"} Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.059269 4818 scope.go:117] "RemoveContainer" containerID="ccfc503d381b73356e74e5dd6e58d14d99f8b895c607615c3330dbd414a9e0f0" Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.059464 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6vkcf" Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.061687 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q95v" event={"ID":"5a1859a2-fa2e-4d8c-bd64-7b63297f258c","Type":"ContainerStarted","Data":"5822a89baddabd937a57882c5fa3bc813379c1149d97c7c1a4e7bd530a67c3bf"} Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.064679 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nchr7" event={"ID":"4b33f4e2-edb8-4e90-b73a-4dc2b13edd41","Type":"ContainerStarted","Data":"e42106a8838c757545b24fa636a233d8e2c877b11dabb87e01cbe3dc7f104fd0"} Dec 03 06:42:49 crc kubenswrapper[4818]: E1203 06:42:49.067927 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q95v" podUID="5a1859a2-fa2e-4d8c-bd64-7b63297f258c" Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.073485 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2d5xh" event={"ID":"e977170c-c8e9-4888-951b-bffa4d4d21d7","Type":"ContainerStarted","Data":"64fb0ab7f6ebb9d72c95c3193b737274cd537b1fa0abd7c708443cf307e0b77f"} Dec 03 06:42:49 crc kubenswrapper[4818]: E1203 06:42:49.081942 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nchr7" podUID="4b33f4e2-edb8-4e90-b73a-4dc2b13edd41" Dec 03 06:42:49 crc kubenswrapper[4818]: E1203 06:42:49.082346 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8hntr" podUID="84dd0135-3ec6-4418-98f9-de8474662893" Dec 03 06:42:49 crc kubenswrapper[4818]: E1203 06:42:49.082440 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jd5hg" podUID="56bb5f70-9dd8-433f-a974-0c95065c1dd1" Dec 03 06:42:49 crc kubenswrapper[4818]: E1203 06:42:49.082485 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bjq7j" podUID="b78bb8f0-22f9-4b8e-93e2-23422db45354" Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.138308 4818 scope.go:117] "RemoveContainer" containerID="5d5943d870e8a32f386d26a15801a4e5e92ab3bb2e26f68513b0031e84f77454" Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.186536 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6vkcf"] Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.190364 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6vkcf"] Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.203429 4818 scope.go:117] "RemoveContainer" containerID="a97f75cf48c42883a2d240e78f3ed84eb6ab09960332d5bc5f75832d2a398586" Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.265145 4818 scope.go:117] "RemoveContainer" containerID="ccfc503d381b73356e74e5dd6e58d14d99f8b895c607615c3330dbd414a9e0f0" Dec 03 06:42:49 crc kubenswrapper[4818]: E1203 06:42:49.265612 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccfc503d381b73356e74e5dd6e58d14d99f8b895c607615c3330dbd414a9e0f0\": container with ID starting with ccfc503d381b73356e74e5dd6e58d14d99f8b895c607615c3330dbd414a9e0f0 not found: ID does not exist" containerID="ccfc503d381b73356e74e5dd6e58d14d99f8b895c607615c3330dbd414a9e0f0" Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.265653 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccfc503d381b73356e74e5dd6e58d14d99f8b895c607615c3330dbd414a9e0f0"} err="failed to get container status \"ccfc503d381b73356e74e5dd6e58d14d99f8b895c607615c3330dbd414a9e0f0\": rpc error: code = NotFound desc = could not find container \"ccfc503d381b73356e74e5dd6e58d14d99f8b895c607615c3330dbd414a9e0f0\": container with ID starting with ccfc503d381b73356e74e5dd6e58d14d99f8b895c607615c3330dbd414a9e0f0 not found: ID does not exist" Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.265681 4818 scope.go:117] "RemoveContainer" containerID="5d5943d870e8a32f386d26a15801a4e5e92ab3bb2e26f68513b0031e84f77454" Dec 03 06:42:49 crc kubenswrapper[4818]: E1203 06:42:49.266012 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d5943d870e8a32f386d26a15801a4e5e92ab3bb2e26f68513b0031e84f77454\": container with ID starting with 5d5943d870e8a32f386d26a15801a4e5e92ab3bb2e26f68513b0031e84f77454 not found: ID does not exist" containerID="5d5943d870e8a32f386d26a15801a4e5e92ab3bb2e26f68513b0031e84f77454" Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.266035 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d5943d870e8a32f386d26a15801a4e5e92ab3bb2e26f68513b0031e84f77454"} err="failed to get container status \"5d5943d870e8a32f386d26a15801a4e5e92ab3bb2e26f68513b0031e84f77454\": rpc error: code = NotFound desc = could not find container \"5d5943d870e8a32f386d26a15801a4e5e92ab3bb2e26f68513b0031e84f77454\": container with ID starting with 5d5943d870e8a32f386d26a15801a4e5e92ab3bb2e26f68513b0031e84f77454 not found: ID does not exist" Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.266048 4818 scope.go:117] "RemoveContainer" containerID="a97f75cf48c42883a2d240e78f3ed84eb6ab09960332d5bc5f75832d2a398586" Dec 03 06:42:49 crc kubenswrapper[4818]: E1203 06:42:49.279726 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a97f75cf48c42883a2d240e78f3ed84eb6ab09960332d5bc5f75832d2a398586\": container with ID starting with a97f75cf48c42883a2d240e78f3ed84eb6ab09960332d5bc5f75832d2a398586 not found: ID does not exist" containerID="a97f75cf48c42883a2d240e78f3ed84eb6ab09960332d5bc5f75832d2a398586" Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.279772 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97f75cf48c42883a2d240e78f3ed84eb6ab09960332d5bc5f75832d2a398586"} err="failed to get container status \"a97f75cf48c42883a2d240e78f3ed84eb6ab09960332d5bc5f75832d2a398586\": rpc error: code = NotFound desc = could not find container \"a97f75cf48c42883a2d240e78f3ed84eb6ab09960332d5bc5f75832d2a398586\": container with ID starting with a97f75cf48c42883a2d240e78f3ed84eb6ab09960332d5bc5f75832d2a398586 not found: ID does not exist" Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.293993 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert\") pod \"infra-operator-controller-manager-57548d458d-gmkzd\" (UID: \"a0b8fddb-4636-4f13-8d0b-3e3220721eca\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" Dec 03 06:42:49 crc kubenswrapper[4818]: E1203 06:42:49.294221 4818 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 06:42:49 crc kubenswrapper[4818]: E1203 06:42:49.294319 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert podName:a0b8fddb-4636-4f13-8d0b-3e3220721eca nodeName:}" failed. No retries permitted until 2025-12-03 06:42:53.294294622 +0000 UTC m=+930.985903374 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert") pod "infra-operator-controller-manager-57548d458d-gmkzd" (UID: "a0b8fddb-4636-4f13-8d0b-3e3220721eca") : secret "infra-operator-webhook-server-cert" not found Dec 03 06:42:49 crc kubenswrapper[4818]: I1203 06:42:49.800900 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd445pgf\" (UID: \"f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" Dec 03 06:42:49 crc kubenswrapper[4818]: E1203 06:42:49.801051 4818 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 06:42:49 crc kubenswrapper[4818]: E1203 06:42:49.801090 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert podName:f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34 nodeName:}" failed. No retries permitted until 2025-12-03 06:42:53.801078628 +0000 UTC m=+931.492687380 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" (UID: "f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 06:42:50 crc kubenswrapper[4818]: E1203 06:42:50.091706 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q95v" podUID="5a1859a2-fa2e-4d8c-bd64-7b63297f258c" Dec 03 06:42:50 crc kubenswrapper[4818]: E1203 06:42:50.094634 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nchr7" podUID="4b33f4e2-edb8-4e90-b73a-4dc2b13edd41" Dec 03 06:42:50 crc kubenswrapper[4818]: I1203 06:42:50.416263 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:42:50 crc kubenswrapper[4818]: E1203 06:42:50.416425 4818 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 06:42:50 crc kubenswrapper[4818]: E1203 06:42:50.418524 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs podName:3a635982-10c6-47cc-8462-74873b98410a nodeName:}" failed. No retries permitted until 2025-12-03 06:42:54.418505678 +0000 UTC m=+932.110114430 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs") pod "openstack-operator-controller-manager-98d87bf44-8h8d6" (UID: "3a635982-10c6-47cc-8462-74873b98410a") : secret "metrics-server-cert" not found Dec 03 06:42:50 crc kubenswrapper[4818]: E1203 06:42:50.418591 4818 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 06:42:50 crc kubenswrapper[4818]: E1203 06:42:50.418667 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs podName:3a635982-10c6-47cc-8462-74873b98410a nodeName:}" failed. No retries permitted until 2025-12-03 06:42:54.418644052 +0000 UTC m=+932.110252804 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs") pod "openstack-operator-controller-manager-98d87bf44-8h8d6" (UID: "3a635982-10c6-47cc-8462-74873b98410a") : secret "webhook-server-cert" not found Dec 03 06:42:50 crc kubenswrapper[4818]: I1203 06:42:50.418825 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:42:50 crc kubenswrapper[4818]: I1203 06:42:50.824590 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b02f78cd-9268-46b0-b66a-717793b48b5d" path="/var/lib/kubelet/pods/b02f78cd-9268-46b0-b66a-717793b48b5d/volumes" Dec 03 06:42:53 crc kubenswrapper[4818]: I1203 06:42:53.321954 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert\") pod \"infra-operator-controller-manager-57548d458d-gmkzd\" (UID: \"a0b8fddb-4636-4f13-8d0b-3e3220721eca\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" Dec 03 06:42:53 crc kubenswrapper[4818]: E1203 06:42:53.322102 4818 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 06:42:53 crc kubenswrapper[4818]: E1203 06:42:53.322341 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert podName:a0b8fddb-4636-4f13-8d0b-3e3220721eca nodeName:}" failed. No retries permitted until 2025-12-03 06:43:01.322326503 +0000 UTC m=+939.013935255 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert") pod "infra-operator-controller-manager-57548d458d-gmkzd" (UID: "a0b8fddb-4636-4f13-8d0b-3e3220721eca") : secret "infra-operator-webhook-server-cert" not found Dec 03 06:42:53 crc kubenswrapper[4818]: I1203 06:42:53.829193 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd445pgf\" (UID: \"f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" Dec 03 06:42:53 crc kubenswrapper[4818]: E1203 06:42:53.829414 4818 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 06:42:53 crc kubenswrapper[4818]: E1203 06:42:53.829471 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert podName:f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34 nodeName:}" failed. No retries permitted until 2025-12-03 06:43:01.829446128 +0000 UTC m=+939.521054880 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" (UID: "f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 06:42:54 crc kubenswrapper[4818]: I1203 06:42:54.438524 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:42:54 crc kubenswrapper[4818]: E1203 06:42:54.439085 4818 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 06:42:54 crc kubenswrapper[4818]: E1203 06:42:54.439205 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs podName:3a635982-10c6-47cc-8462-74873b98410a nodeName:}" failed. No retries permitted until 2025-12-03 06:43:02.439172137 +0000 UTC m=+940.130780889 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs") pod "openstack-operator-controller-manager-98d87bf44-8h8d6" (UID: "3a635982-10c6-47cc-8462-74873b98410a") : secret "webhook-server-cert" not found Dec 03 06:42:54 crc kubenswrapper[4818]: I1203 06:42:54.439809 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:42:54 crc kubenswrapper[4818]: E1203 06:42:54.439951 4818 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 06:42:54 crc kubenswrapper[4818]: E1203 06:42:54.440033 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs podName:3a635982-10c6-47cc-8462-74873b98410a nodeName:}" failed. No retries permitted until 2025-12-03 06:43:02.440020948 +0000 UTC m=+940.131629770 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs") pod "openstack-operator-controller-manager-98d87bf44-8h8d6" (UID: "3a635982-10c6-47cc-8462-74873b98410a") : secret "metrics-server-cert" not found Dec 03 06:42:57 crc kubenswrapper[4818]: I1203 06:42:57.176749 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bqx2t"] Dec 03 06:42:57 crc kubenswrapper[4818]: E1203 06:42:57.177037 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b02f78cd-9268-46b0-b66a-717793b48b5d" containerName="extract-utilities" Dec 03 06:42:57 crc kubenswrapper[4818]: I1203 06:42:57.177049 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b02f78cd-9268-46b0-b66a-717793b48b5d" containerName="extract-utilities" Dec 03 06:42:57 crc kubenswrapper[4818]: E1203 06:42:57.177060 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b02f78cd-9268-46b0-b66a-717793b48b5d" containerName="registry-server" Dec 03 06:42:57 crc kubenswrapper[4818]: I1203 06:42:57.177067 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b02f78cd-9268-46b0-b66a-717793b48b5d" containerName="registry-server" Dec 03 06:42:57 crc kubenswrapper[4818]: E1203 06:42:57.177078 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b02f78cd-9268-46b0-b66a-717793b48b5d" containerName="extract-content" Dec 03 06:42:57 crc kubenswrapper[4818]: I1203 06:42:57.177084 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b02f78cd-9268-46b0-b66a-717793b48b5d" containerName="extract-content" Dec 03 06:42:57 crc kubenswrapper[4818]: I1203 06:42:57.177213 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="b02f78cd-9268-46b0-b66a-717793b48b5d" containerName="registry-server" Dec 03 06:42:57 crc kubenswrapper[4818]: I1203 06:42:57.178248 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bqx2t" Dec 03 06:42:57 crc kubenswrapper[4818]: I1203 06:42:57.194729 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bqx2t"] Dec 03 06:42:57 crc kubenswrapper[4818]: I1203 06:42:57.326002 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-utilities\") pod \"certified-operators-bqx2t\" (UID: \"f4158ae0-a176-49a2-a8b8-57cd52efbdc3\") " pod="openshift-marketplace/certified-operators-bqx2t" Dec 03 06:42:57 crc kubenswrapper[4818]: I1203 06:42:57.326392 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-catalog-content\") pod \"certified-operators-bqx2t\" (UID: \"f4158ae0-a176-49a2-a8b8-57cd52efbdc3\") " pod="openshift-marketplace/certified-operators-bqx2t" Dec 03 06:42:57 crc kubenswrapper[4818]: I1203 06:42:57.326494 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcjpk\" (UniqueName: \"kubernetes.io/projected/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-kube-api-access-gcjpk\") pod \"certified-operators-bqx2t\" (UID: \"f4158ae0-a176-49a2-a8b8-57cd52efbdc3\") " pod="openshift-marketplace/certified-operators-bqx2t" Dec 03 06:42:57 crc kubenswrapper[4818]: I1203 06:42:57.428940 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-utilities\") pod \"certified-operators-bqx2t\" (UID: \"f4158ae0-a176-49a2-a8b8-57cd52efbdc3\") " pod="openshift-marketplace/certified-operators-bqx2t" Dec 03 06:42:57 crc kubenswrapper[4818]: I1203 06:42:57.428983 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-catalog-content\") pod \"certified-operators-bqx2t\" (UID: \"f4158ae0-a176-49a2-a8b8-57cd52efbdc3\") " pod="openshift-marketplace/certified-operators-bqx2t" Dec 03 06:42:57 crc kubenswrapper[4818]: I1203 06:42:57.429033 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcjpk\" (UniqueName: \"kubernetes.io/projected/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-kube-api-access-gcjpk\") pod \"certified-operators-bqx2t\" (UID: \"f4158ae0-a176-49a2-a8b8-57cd52efbdc3\") " pod="openshift-marketplace/certified-operators-bqx2t" Dec 03 06:42:57 crc kubenswrapper[4818]: I1203 06:42:57.429647 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-utilities\") pod \"certified-operators-bqx2t\" (UID: \"f4158ae0-a176-49a2-a8b8-57cd52efbdc3\") " pod="openshift-marketplace/certified-operators-bqx2t" Dec 03 06:42:57 crc kubenswrapper[4818]: I1203 06:42:57.429717 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-catalog-content\") pod \"certified-operators-bqx2t\" (UID: \"f4158ae0-a176-49a2-a8b8-57cd52efbdc3\") " pod="openshift-marketplace/certified-operators-bqx2t" Dec 03 06:42:57 crc kubenswrapper[4818]: I1203 06:42:57.456683 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcjpk\" (UniqueName: \"kubernetes.io/projected/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-kube-api-access-gcjpk\") pod \"certified-operators-bqx2t\" (UID: \"f4158ae0-a176-49a2-a8b8-57cd52efbdc3\") " pod="openshift-marketplace/certified-operators-bqx2t" Dec 03 06:42:57 crc kubenswrapper[4818]: I1203 06:42:57.499140 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bqx2t" Dec 03 06:43:01 crc kubenswrapper[4818]: I1203 06:43:01.372276 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert\") pod \"infra-operator-controller-manager-57548d458d-gmkzd\" (UID: \"a0b8fddb-4636-4f13-8d0b-3e3220721eca\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" Dec 03 06:43:01 crc kubenswrapper[4818]: I1203 06:43:01.377449 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0b8fddb-4636-4f13-8d0b-3e3220721eca-cert\") pod \"infra-operator-controller-manager-57548d458d-gmkzd\" (UID: \"a0b8fddb-4636-4f13-8d0b-3e3220721eca\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" Dec 03 06:43:01 crc kubenswrapper[4818]: I1203 06:43:01.647393 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" Dec 03 06:43:01 crc kubenswrapper[4818]: I1203 06:43:01.878844 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd445pgf\" (UID: \"f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" Dec 03 06:43:01 crc kubenswrapper[4818]: I1203 06:43:01.883234 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd445pgf\" (UID: \"f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" Dec 03 06:43:02 crc kubenswrapper[4818]: I1203 06:43:02.095909 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" Dec 03 06:43:02 crc kubenswrapper[4818]: E1203 06:43:02.433920 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801" Dec 03 06:43:02 crc kubenswrapper[4818]: E1203 06:43:02.434075 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wv8p5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-4hdr6_openstack-operators(4130e7e0-4596-451f-b346-5afba08e9dea): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:43:02 crc kubenswrapper[4818]: I1203 06:43:02.520552 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:43:02 crc kubenswrapper[4818]: I1203 06:43:02.520656 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:43:02 crc kubenswrapper[4818]: I1203 06:43:02.524922 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-webhook-certs\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:43:02 crc kubenswrapper[4818]: I1203 06:43:02.526038 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a635982-10c6-47cc-8462-74873b98410a-metrics-certs\") pod \"openstack-operator-controller-manager-98d87bf44-8h8d6\" (UID: \"3a635982-10c6-47cc-8462-74873b98410a\") " pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:43:02 crc kubenswrapper[4818]: I1203 06:43:02.620148 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:43:09 crc kubenswrapper[4818]: E1203 06:43:09.327207 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530" Dec 03 06:43:09 crc kubenswrapper[4818]: E1203 06:43:09.327782 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zj878,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-hpnhz_openstack-operators(3d1565f0-62d4-4ecd-9d73-16338035196e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:43:10 crc kubenswrapper[4818]: E1203 06:43:10.689631 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 03 06:43:10 crc kubenswrapper[4818]: E1203 06:43:10.690390 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-278sm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-thzjt_openstack-operators(6d5bf0d5-df89-4ab4-a711-e6f7f3d1123e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:43:11 crc kubenswrapper[4818]: E1203 06:43:11.405160 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7" Dec 03 06:43:11 crc kubenswrapper[4818]: E1203 06:43:11.405396 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jgpfh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-98z98_openstack-operators(df292cb6-9f13-4cec-a7e1-1ef3155a7b1f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:43:12 crc kubenswrapper[4818]: E1203 06:43:12.005963 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809" Dec 03 06:43:12 crc kubenswrapper[4818]: E1203 06:43:12.006140 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zx6wv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-2qvsd_openstack-operators(528c7c89-7980-4acc-9321-ff56c9d38489): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:43:12 crc kubenswrapper[4818]: E1203 06:43:12.916982 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85" Dec 03 06:43:12 crc kubenswrapper[4818]: E1203 06:43:12.917223 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ncrl8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-wndt9_openstack-operators(91e645ae-28ec-4d3b-8fe3-e75aa1acf5a5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:43:13 crc kubenswrapper[4818]: I1203 06:43:13.302144 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:43:13 crc kubenswrapper[4818]: I1203 06:43:13.302224 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:43:13 crc kubenswrapper[4818]: E1203 06:43:13.505964 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 03 06:43:13 crc kubenswrapper[4818]: E1203 06:43:13.506155 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gvqj7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-xtq4s_openstack-operators(1dc6c089-07b6-4818-a138-6b6a8c0cd1b5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:43:16 crc kubenswrapper[4818]: E1203 06:43:16.900946 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 03 06:43:16 crc kubenswrapper[4818]: E1203 06:43:16.901696 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fcc48,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-27nrr_openstack-operators(1bd3fb88-7598-4a28-90fe-324d855dc90e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:43:18 crc kubenswrapper[4818]: I1203 06:43:18.108261 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd"] Dec 03 06:43:18 crc kubenswrapper[4818]: I1203 06:43:18.120093 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf"] Dec 03 06:43:18 crc kubenswrapper[4818]: I1203 06:43:18.261980 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6"] Dec 03 06:43:18 crc kubenswrapper[4818]: I1203 06:43:18.352288 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bqx2t"] Dec 03 06:43:18 crc kubenswrapper[4818]: W1203 06:43:18.389157 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a635982_10c6_47cc_8462_74873b98410a.slice/crio-b898a285b8ff82bfd858abef48dd7f42e8be1d4ccd7b21008fa0d9407d5a8c6c WatchSource:0}: Error finding container b898a285b8ff82bfd858abef48dd7f42e8be1d4ccd7b21008fa0d9407d5a8c6c: Status 404 returned error can't find the container with id b898a285b8ff82bfd858abef48dd7f42e8be1d4ccd7b21008fa0d9407d5a8c6c Dec 03 06:43:18 crc kubenswrapper[4818]: W1203 06:43:18.392011 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0b8fddb_4636_4f13_8d0b_3e3220721eca.slice/crio-50e2ccebd04785f3fdeaeebe7e0567d1a8f2605105b31b0f80b40ebef92f6baf WatchSource:0}: Error finding container 50e2ccebd04785f3fdeaeebe7e0567d1a8f2605105b31b0f80b40ebef92f6baf: Status 404 returned error can't find the container with id 50e2ccebd04785f3fdeaeebe7e0567d1a8f2605105b31b0f80b40ebef92f6baf Dec 03 06:43:18 crc kubenswrapper[4818]: W1203 06:43:18.394188 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf24eeb0d_5a36_4dd7_93e8_3e87f1fa7e34.slice/crio-26a01b3d04ad2b84e46a21ac3cb2e0385b4659318179af0448e204fe4dd7d201 WatchSource:0}: Error finding container 26a01b3d04ad2b84e46a21ac3cb2e0385b4659318179af0448e204fe4dd7d201: Status 404 returned error can't find the container with id 26a01b3d04ad2b84e46a21ac3cb2e0385b4659318179af0448e204fe4dd7d201 Dec 03 06:43:18 crc kubenswrapper[4818]: W1203 06:43:18.417661 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4158ae0_a176_49a2_a8b8_57cd52efbdc3.slice/crio-7653fa9f81f80003eb880fe94a15686d2220bf1217884afa2d4f482990d0fa26 WatchSource:0}: Error finding container 7653fa9f81f80003eb880fe94a15686d2220bf1217884afa2d4f482990d0fa26: Status 404 returned error can't find the container with id 7653fa9f81f80003eb880fe94a15686d2220bf1217884afa2d4f482990d0fa26 Dec 03 06:43:18 crc kubenswrapper[4818]: I1203 06:43:18.461428 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" event={"ID":"f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34","Type":"ContainerStarted","Data":"26a01b3d04ad2b84e46a21ac3cb2e0385b4659318179af0448e204fe4dd7d201"} Dec 03 06:43:18 crc kubenswrapper[4818]: I1203 06:43:18.462686 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bqx2t" event={"ID":"f4158ae0-a176-49a2-a8b8-57cd52efbdc3","Type":"ContainerStarted","Data":"7653fa9f81f80003eb880fe94a15686d2220bf1217884afa2d4f482990d0fa26"} Dec 03 06:43:18 crc kubenswrapper[4818]: I1203 06:43:18.464031 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" event={"ID":"3a635982-10c6-47cc-8462-74873b98410a","Type":"ContainerStarted","Data":"b898a285b8ff82bfd858abef48dd7f42e8be1d4ccd7b21008fa0d9407d5a8c6c"} Dec 03 06:43:18 crc kubenswrapper[4818]: I1203 06:43:18.465183 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" event={"ID":"a0b8fddb-4636-4f13-8d0b-3e3220721eca","Type":"ContainerStarted","Data":"50e2ccebd04785f3fdeaeebe7e0567d1a8f2605105b31b0f80b40ebef92f6baf"} Dec 03 06:43:19 crc kubenswrapper[4818]: I1203 06:43:19.492531 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-z6c4f" event={"ID":"7c36b14e-4e1d-4956-a505-ecdffb9d9f9c","Type":"ContainerStarted","Data":"b25090b3ad3329bb76efc45ad20db7cbf2f5c48951d2b2b7a2599b617d70cc97"} Dec 03 06:43:19 crc kubenswrapper[4818]: I1203 06:43:19.539974 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gtmmz" event={"ID":"b29acffb-a7e4-4029-93fa-20f20de27dfe","Type":"ContainerStarted","Data":"2bd544c22015c50867f84a1f3f322468d038a7389aaf483d81fd9ae86ce1d32d"} Dec 03 06:43:19 crc kubenswrapper[4818]: I1203 06:43:19.543644 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2d5xh" event={"ID":"e977170c-c8e9-4888-951b-bffa4d4d21d7","Type":"ContainerStarted","Data":"8e468184eca0b6e70d1d1bba1307f58bdd1daef424643c22107284677eeddf52"} Dec 03 06:43:19 crc kubenswrapper[4818]: I1203 06:43:19.544910 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-6rr64" event={"ID":"45dcf421-464d-4eb9-a23f-9924ecf5bb03","Type":"ContainerStarted","Data":"c14b45dd60d0fbb580ea5e7d7e0081f96b03a2c220062d9de5975f612d996662"} Dec 03 06:43:19 crc kubenswrapper[4818]: I1203 06:43:19.546183 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b4thx" event={"ID":"0d4b3be5-bb2e-4580-8142-60d15b16c36f","Type":"ContainerStarted","Data":"8823beaeda2c014363063e320c951ac7472f0f9c15a71295aadf4b54b1540499"} Dec 03 06:43:19 crc kubenswrapper[4818]: I1203 06:43:19.547760 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-s84mt" event={"ID":"0c360185-7fff-4054-a85e-ca2bbb1fa5e0","Type":"ContainerStarted","Data":"66339f5d7ff63a15daa108f5c50b671d3b98b7d7f25d4b173c6be63cc3522b30"} Dec 03 06:43:19 crc kubenswrapper[4818]: I1203 06:43:19.549796 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jd5hg" event={"ID":"56bb5f70-9dd8-433f-a974-0c95065c1dd1","Type":"ContainerStarted","Data":"01296a99fc1bdd51fd40bd64c06308dfdc4112febe7cc92df29a31b6c2eccfd2"} Dec 03 06:43:19 crc kubenswrapper[4818]: I1203 06:43:19.551049 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bjq7j" event={"ID":"b78bb8f0-22f9-4b8e-93e2-23422db45354","Type":"ContainerStarted","Data":"6c3f5d29ef9f9655af05c7b6b778734bce4e0335036b369f7bd594194dc69bf9"} Dec 03 06:43:19 crc kubenswrapper[4818]: I1203 06:43:19.554011 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-r4b9n" event={"ID":"2e8a3376-fa5d-458d-a659-0c3cef8af10b","Type":"ContainerStarted","Data":"4c616aaa46a2ef4109d9c3ddf5c538c90c6aa27921fe4afc0ccf12ba9f42e0e3"} Dec 03 06:43:20 crc kubenswrapper[4818]: I1203 06:43:20.580000 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8hntr" event={"ID":"84dd0135-3ec6-4418-98f9-de8474662893","Type":"ContainerStarted","Data":"0ba4251f1258a2f645e5760f3812eaff72bba01f5726bb2ac1fd1a44cafe99bc"} Dec 03 06:43:20 crc kubenswrapper[4818]: I1203 06:43:20.595135 4818 generic.go:334] "Generic (PLEG): container finished" podID="f4158ae0-a176-49a2-a8b8-57cd52efbdc3" containerID="4c54f8a69a25fe4acc31175f5a5c6e830aaadb6ae44469cbdba5fa348eabf056" exitCode=0 Dec 03 06:43:20 crc kubenswrapper[4818]: I1203 06:43:20.595234 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bqx2t" event={"ID":"f4158ae0-a176-49a2-a8b8-57cd52efbdc3","Type":"ContainerDied","Data":"4c54f8a69a25fe4acc31175f5a5c6e830aaadb6ae44469cbdba5fa348eabf056"} Dec 03 06:43:20 crc kubenswrapper[4818]: I1203 06:43:20.601049 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" event={"ID":"3a635982-10c6-47cc-8462-74873b98410a","Type":"ContainerStarted","Data":"eea28dc438924bbb297c449835f7d0f2c5028b1f79a0cc1d6bd59608746aea4c"} Dec 03 06:43:20 crc kubenswrapper[4818]: I1203 06:43:20.601714 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:43:20 crc kubenswrapper[4818]: I1203 06:43:20.711941 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nchr7" event={"ID":"4b33f4e2-edb8-4e90-b73a-4dc2b13edd41","Type":"ContainerStarted","Data":"5a1ab7aad57b42c8f7ce22af2396a215d8e07823f5ec7f98194cac3b4eb0897a"} Dec 03 06:43:22 crc kubenswrapper[4818]: I1203 06:43:22.770110 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" podStartSLOduration=36.770092838 podStartE2EDuration="36.770092838s" podCreationTimestamp="2025-12-03 06:42:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:43:20.752787176 +0000 UTC m=+958.444395928" watchObservedRunningTime="2025-12-03 06:43:22.770092838 +0000 UTC m=+960.461701590" Dec 03 06:43:23 crc kubenswrapper[4818]: I1203 06:43:23.742594 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q95v" event={"ID":"5a1859a2-fa2e-4d8c-bd64-7b63297f258c","Type":"ContainerStarted","Data":"8eed1925c4eb277bb982ba2995150f5260c79a42ecf4a9cb5f77776eecafea46"} Dec 03 06:43:23 crc kubenswrapper[4818]: I1203 06:43:23.760267 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5q95v" podStartSLOduration=7.259316981 podStartE2EDuration="37.76024679s" podCreationTimestamp="2025-12-03 06:42:46 +0000 UTC" firstStartedPulling="2025-12-03 06:42:48.067630965 +0000 UTC m=+925.759239717" lastFinishedPulling="2025-12-03 06:43:18.568560774 +0000 UTC m=+956.260169526" observedRunningTime="2025-12-03 06:43:23.756772844 +0000 UTC m=+961.448381626" watchObservedRunningTime="2025-12-03 06:43:23.76024679 +0000 UTC m=+961.451855542" Dec 03 06:43:24 crc kubenswrapper[4818]: E1203 06:43:24.164765 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-4hdr6" podUID="4130e7e0-4596-451f-b346-5afba08e9dea" Dec 03 06:43:24 crc kubenswrapper[4818]: E1203 06:43:24.209718 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-27nrr" podUID="1bd3fb88-7598-4a28-90fe-324d855dc90e" Dec 03 06:43:24 crc kubenswrapper[4818]: E1203 06:43:24.237994 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-thzjt" podUID="6d5bf0d5-df89-4ab4-a711-e6f7f3d1123e" Dec 03 06:43:24 crc kubenswrapper[4818]: E1203 06:43:24.272468 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpnhz" podUID="3d1565f0-62d4-4ecd-9d73-16338035196e" Dec 03 06:43:24 crc kubenswrapper[4818]: E1203 06:43:24.453229 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xtq4s" podUID="1dc6c089-07b6-4818-a138-6b6a8c0cd1b5" Dec 03 06:43:24 crc kubenswrapper[4818]: E1203 06:43:24.504485 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wndt9" podUID="91e645ae-28ec-4d3b-8fe3-e75aa1acf5a5" Dec 03 06:43:24 crc kubenswrapper[4818]: E1203 06:43:24.572504 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-2qvsd" podUID="528c7c89-7980-4acc-9321-ff56c9d38489" Dec 03 06:43:24 crc kubenswrapper[4818]: E1203 06:43:24.651234 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-98z98" podUID="df292cb6-9f13-4cec-a7e1-1ef3155a7b1f" Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.751997 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-2qvsd" event={"ID":"528c7c89-7980-4acc-9321-ff56c9d38489","Type":"ContainerStarted","Data":"d09a1e8b424a47975dc75fd0567709e85989ae964e75d7847926fe0998870b34"} Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.766086 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bjq7j" event={"ID":"b78bb8f0-22f9-4b8e-93e2-23422db45354","Type":"ContainerStarted","Data":"c976a1260a65e1fc618afb55b98fdf2aec366e1519947372d204082c1ee84319"} Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.766796 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bjq7j" Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.773386 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bjq7j" Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.774373 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-27nrr" event={"ID":"1bd3fb88-7598-4a28-90fe-324d855dc90e","Type":"ContainerStarted","Data":"710faaee62ab027fadf2d6ae5ad11b4f22e7abc165a9df7fa840437aa45ac82f"} Dec 03 06:43:24 crc kubenswrapper[4818]: E1203 06:43:24.776694 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-27nrr" podUID="1bd3fb88-7598-4a28-90fe-324d855dc90e" Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.779327 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" event={"ID":"f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34","Type":"ContainerStarted","Data":"3eeb0abc40d5c24f07354eb85f3a7458e904aea9c526b07b730244bf73616ae1"} Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.782681 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" event={"ID":"a0b8fddb-4636-4f13-8d0b-3e3220721eca","Type":"ContainerStarted","Data":"28c58caf31cd62f399bbbb9c54e72094d69311a1ea2fad10c5705a60db8951b9"} Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.787008 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-thzjt" event={"ID":"6d5bf0d5-df89-4ab4-a711-e6f7f3d1123e","Type":"ContainerStarted","Data":"34accb4332bcfd96907e3dfb2d26c5df00c2631803f82875d961f8906341e4ac"} Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.798689 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpnhz" event={"ID":"3d1565f0-62d4-4ecd-9d73-16338035196e","Type":"ContainerStarted","Data":"bba2075ca0468ed167a3077df518b5b3662f67b44934d14f97cd10e07982cc64"} Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.817116 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-6rr64" event={"ID":"45dcf421-464d-4eb9-a23f-9924ecf5bb03","Type":"ContainerStarted","Data":"0b9880a291da80c51453f8bb984f2f1d4ac1539a2f3ef95aa406c159dda31d7b"} Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.817883 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-6rr64" Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.820665 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-6rr64" Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.838111 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xtq4s" event={"ID":"1dc6c089-07b6-4818-a138-6b6a8c0cd1b5","Type":"ContainerStarted","Data":"46939b2035261567ff5606d2121882ae12938fda01751c0222924ea98fba6af1"} Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.845934 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bjq7j" podStartSLOduration=3.935966827 podStartE2EDuration="39.845916532s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:47.919233715 +0000 UTC m=+925.610842457" lastFinishedPulling="2025-12-03 06:43:23.82918341 +0000 UTC m=+961.520792162" observedRunningTime="2025-12-03 06:43:24.842266731 +0000 UTC m=+962.533875483" watchObservedRunningTime="2025-12-03 06:43:24.845916532 +0000 UTC m=+962.537525284" Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.864178 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-98z98" event={"ID":"df292cb6-9f13-4cec-a7e1-1ef3155a7b1f","Type":"ContainerStarted","Data":"c71d36ac91b8bb561c36def29da4800018796e640b7b1c004d45d3937b750d1a"} Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.879588 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-s84mt" event={"ID":"0c360185-7fff-4054-a85e-ca2bbb1fa5e0","Type":"ContainerStarted","Data":"e8370de8ae19271213c98497a2d7eec79909103fee8ce1be247f343cc0989dc9"} Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.879830 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-s84mt" Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.886029 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-s84mt" Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.901071 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wndt9" event={"ID":"91e645ae-28ec-4d3b-8fe3-e75aa1acf5a5","Type":"ContainerStarted","Data":"7d30676c46acca2030b8d7890a47bef848d39690299966e36099efa09ae39a99"} Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.926674 4818 generic.go:334] "Generic (PLEG): container finished" podID="f4158ae0-a176-49a2-a8b8-57cd52efbdc3" containerID="af7698548d1e27d130f3deb9798775769cbde26be79f442ec2ca68667ca15b6a" exitCode=0 Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.926771 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bqx2t" event={"ID":"f4158ae0-a176-49a2-a8b8-57cd52efbdc3","Type":"ContainerDied","Data":"af7698548d1e27d130f3deb9798775769cbde26be79f442ec2ca68667ca15b6a"} Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.944283 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-4hdr6" event={"ID":"4130e7e0-4596-451f-b346-5afba08e9dea","Type":"ContainerStarted","Data":"037633ee31a7a5215f8e52c2032f50bfb9ca88a00ebdd962d1f9cb9f1d1cbc8b"} Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.978419 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-r4b9n" event={"ID":"2e8a3376-fa5d-458d-a659-0c3cef8af10b","Type":"ContainerStarted","Data":"1117f769f00d70dfa4441be18842ce09aebb5fcf7d3fbfaa0244e8c19d3168d0"} Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.979162 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-r4b9n" Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.979770 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-6rr64" podStartSLOduration=3.889443034 podStartE2EDuration="39.979759751s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:47.757518935 +0000 UTC m=+925.449127687" lastFinishedPulling="2025-12-03 06:43:23.847835652 +0000 UTC m=+961.539444404" observedRunningTime="2025-12-03 06:43:24.977226718 +0000 UTC m=+962.668835470" watchObservedRunningTime="2025-12-03 06:43:24.979759751 +0000 UTC m=+962.671368503" Dec 03 06:43:24 crc kubenswrapper[4818]: I1203 06:43:24.983909 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-r4b9n" Dec 03 06:43:25 crc kubenswrapper[4818]: I1203 06:43:25.100515 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-s84mt" podStartSLOduration=3.9280203 podStartE2EDuration="40.100498504s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:47.742311898 +0000 UTC m=+925.433920650" lastFinishedPulling="2025-12-03 06:43:23.914790102 +0000 UTC m=+961.606398854" observedRunningTime="2025-12-03 06:43:25.095034679 +0000 UTC m=+962.786643441" watchObservedRunningTime="2025-12-03 06:43:25.100498504 +0000 UTC m=+962.792107256" Dec 03 06:43:25 crc kubenswrapper[4818]: I1203 06:43:25.175353 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-r4b9n" podStartSLOduration=4.260578107 podStartE2EDuration="40.17533047s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:47.899014494 +0000 UTC m=+925.590623246" lastFinishedPulling="2025-12-03 06:43:23.813766847 +0000 UTC m=+961.505375609" observedRunningTime="2025-12-03 06:43:25.155269123 +0000 UTC m=+962.846877885" watchObservedRunningTime="2025-12-03 06:43:25.17533047 +0000 UTC m=+962.866939222" Dec 03 06:43:25 crc kubenswrapper[4818]: I1203 06:43:25.990043 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" event={"ID":"a0b8fddb-4636-4f13-8d0b-3e3220721eca","Type":"ContainerStarted","Data":"f4a873a64269e0dd7395b8527332b323a81639cbeba5187ea2d2b8053143a93b"} Dec 03 06:43:25 crc kubenswrapper[4818]: I1203 06:43:25.990347 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" Dec 03 06:43:25 crc kubenswrapper[4818]: I1203 06:43:25.991976 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-2qvsd" event={"ID":"528c7c89-7980-4acc-9321-ff56c9d38489","Type":"ContainerStarted","Data":"30ed49b85d0fbf76805aaa948c59a42b822797b058d481bca2974db78070bad0"} Dec 03 06:43:25 crc kubenswrapper[4818]: I1203 06:43:25.992034 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-2qvsd" Dec 03 06:43:25 crc kubenswrapper[4818]: I1203 06:43:25.998628 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b4thx" event={"ID":"0d4b3be5-bb2e-4580-8142-60d15b16c36f","Type":"ContainerStarted","Data":"d126b51d45e291fd4c8a2fc1050ec893e852e5a58dce14a522eaa0c6004b8fed"} Dec 03 06:43:25 crc kubenswrapper[4818]: I1203 06:43:25.999386 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b4thx" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.001132 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b4thx" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.007046 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" event={"ID":"f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34","Type":"ContainerStarted","Data":"16dc7bf026d4a3e81caf292ee8af5b369f8603357a0a800a52ea5249c9c226ff"} Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.007600 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.010932 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jd5hg" event={"ID":"56bb5f70-9dd8-433f-a974-0c95065c1dd1","Type":"ContainerStarted","Data":"4378f5fe91fe898047b7adba629ada2ae3fea7e9cc04db48099ae26303aaa1af"} Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.011674 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jd5hg" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.015107 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-z6c4f" event={"ID":"7c36b14e-4e1d-4956-a505-ecdffb9d9f9c","Type":"ContainerStarted","Data":"66d975217a611cb595a750a4a27b3844c183f56f5acde9d03743a34ecf2af5f7"} Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.015512 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-z6c4f" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.015681 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jd5hg" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.018028 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" podStartSLOduration=35.700281343 podStartE2EDuration="41.018014105s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:43:18.394932108 +0000 UTC m=+956.086540860" lastFinishedPulling="2025-12-03 06:43:23.71266486 +0000 UTC m=+961.404273622" observedRunningTime="2025-12-03 06:43:26.01336325 +0000 UTC m=+963.704972002" watchObservedRunningTime="2025-12-03 06:43:26.018014105 +0000 UTC m=+963.709622857" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.020315 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-z6c4f" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.028755 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nchr7" event={"ID":"4b33f4e2-edb8-4e90-b73a-4dc2b13edd41","Type":"ContainerStarted","Data":"ae0819a9628bd60202f4fd85ffef8162981346e71d5077bb9ad45362539d3620"} Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.029807 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nchr7" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.035250 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nchr7" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.041621 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gtmmz" event={"ID":"b29acffb-a7e4-4029-93fa-20f20de27dfe","Type":"ContainerStarted","Data":"fbc360e00b628db3f3bf033e026430d53ce7830086a556ac8a571695d5b0a883"} Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.042129 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gtmmz" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.048179 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gtmmz" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.053111 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jd5hg" podStartSLOduration=5.101802705 podStartE2EDuration="41.053088535s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:47.917613825 +0000 UTC m=+925.609222567" lastFinishedPulling="2025-12-03 06:43:23.868899645 +0000 UTC m=+961.560508397" observedRunningTime="2025-12-03 06:43:26.04161171 +0000 UTC m=+963.733220462" watchObservedRunningTime="2025-12-03 06:43:26.053088535 +0000 UTC m=+963.744697287" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.055002 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2d5xh" event={"ID":"e977170c-c8e9-4888-951b-bffa4d4d21d7","Type":"ContainerStarted","Data":"5a94630f31050d62f4da0284277202be70a5a2ff3086e8fcbcb98620d0e5224a"} Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.055895 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2d5xh" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.093513 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xtq4s" event={"ID":"1dc6c089-07b6-4818-a138-6b6a8c0cd1b5","Type":"ContainerStarted","Data":"e4ffb6749dfb6c8fb1f466062d2766e0528d196e5ce8335113fa98b1b7ae861b"} Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.094291 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xtq4s" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.098112 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" podStartSLOduration=35.739324881 podStartE2EDuration="41.098094921s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:43:18.396352803 +0000 UTC m=+956.087961555" lastFinishedPulling="2025-12-03 06:43:23.755122833 +0000 UTC m=+961.446731595" observedRunningTime="2025-12-03 06:43:26.094207584 +0000 UTC m=+963.785816336" watchObservedRunningTime="2025-12-03 06:43:26.098094921 +0000 UTC m=+963.789703673" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.103678 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2d5xh" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.127418 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-thzjt" event={"ID":"6d5bf0d5-df89-4ab4-a711-e6f7f3d1123e","Type":"ContainerStarted","Data":"c26ac53d3ccfd9400ee0ed7954ad84a6773e5e032d044c61d92589b241790fa6"} Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.128993 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-thzjt" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.135452 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b4thx" podStartSLOduration=5.045066888 podStartE2EDuration="41.135432337s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:47.746070421 +0000 UTC m=+925.437679173" lastFinishedPulling="2025-12-03 06:43:23.83643587 +0000 UTC m=+961.528044622" observedRunningTime="2025-12-03 06:43:26.133212222 +0000 UTC m=+963.824820974" watchObservedRunningTime="2025-12-03 06:43:26.135432337 +0000 UTC m=+963.827041089" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.147774 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8hntr" event={"ID":"84dd0135-3ec6-4418-98f9-de8474662893","Type":"ContainerStarted","Data":"08f96ee26d16a03d7b1beb279809604c1b5750317ff327e191b77c8e28cb3cce"} Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.147809 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8hntr" Dec 03 06:43:26 crc kubenswrapper[4818]: E1203 06:43:26.149570 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-27nrr" podUID="1bd3fb88-7598-4a28-90fe-324d855dc90e" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.153113 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8hntr" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.160075 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-2qvsd" podStartSLOduration=3.677483617 podStartE2EDuration="41.160055197s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:47.758638973 +0000 UTC m=+925.450247725" lastFinishedPulling="2025-12-03 06:43:25.241210553 +0000 UTC m=+962.932819305" observedRunningTime="2025-12-03 06:43:26.157049743 +0000 UTC m=+963.848658495" watchObservedRunningTime="2025-12-03 06:43:26.160055197 +0000 UTC m=+963.851663969" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.201481 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-thzjt" podStartSLOduration=3.320357802 podStartE2EDuration="41.201461844s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:47.523259966 +0000 UTC m=+925.214868718" lastFinishedPulling="2025-12-03 06:43:25.404364008 +0000 UTC m=+963.095972760" observedRunningTime="2025-12-03 06:43:26.198233414 +0000 UTC m=+963.889842166" watchObservedRunningTime="2025-12-03 06:43:26.201461844 +0000 UTC m=+963.893070596" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.245889 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-2d5xh" podStartSLOduration=5.384838264 podStartE2EDuration="41.245873295s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:48.029672534 +0000 UTC m=+925.721281276" lastFinishedPulling="2025-12-03 06:43:23.890707555 +0000 UTC m=+961.582316307" observedRunningTime="2025-12-03 06:43:26.245227109 +0000 UTC m=+963.936835861" watchObservedRunningTime="2025-12-03 06:43:26.245873295 +0000 UTC m=+963.937482047" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.276926 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xtq4s" podStartSLOduration=3.711512431 podStartE2EDuration="41.276908615s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:47.730622998 +0000 UTC m=+925.422231750" lastFinishedPulling="2025-12-03 06:43:25.296019182 +0000 UTC m=+962.987627934" observedRunningTime="2025-12-03 06:43:26.276175497 +0000 UTC m=+963.967784249" watchObservedRunningTime="2025-12-03 06:43:26.276908615 +0000 UTC m=+963.968517367" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.364024 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-z6c4f" podStartSLOduration=4.517234109 podStartE2EDuration="41.364003654s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:46.987022139 +0000 UTC m=+924.678630891" lastFinishedPulling="2025-12-03 06:43:23.833791684 +0000 UTC m=+961.525400436" observedRunningTime="2025-12-03 06:43:26.322351742 +0000 UTC m=+964.013960494" watchObservedRunningTime="2025-12-03 06:43:26.364003654 +0000 UTC m=+964.055612406" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.365596 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-8hntr" podStartSLOduration=5.391202652 podStartE2EDuration="41.365587944s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:47.917162654 +0000 UTC m=+925.608771406" lastFinishedPulling="2025-12-03 06:43:23.891547946 +0000 UTC m=+961.583156698" observedRunningTime="2025-12-03 06:43:26.362955719 +0000 UTC m=+964.054564471" watchObservedRunningTime="2025-12-03 06:43:26.365587944 +0000 UTC m=+964.057196686" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.388790 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gtmmz" podStartSLOduration=5.025469783 podStartE2EDuration="41.388773409s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:47.472722203 +0000 UTC m=+925.164330955" lastFinishedPulling="2025-12-03 06:43:23.836025819 +0000 UTC m=+961.527634581" observedRunningTime="2025-12-03 06:43:26.386702027 +0000 UTC m=+964.078310779" watchObservedRunningTime="2025-12-03 06:43:26.388773409 +0000 UTC m=+964.080382161" Dec 03 06:43:26 crc kubenswrapper[4818]: I1203 06:43:26.442730 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-nchr7" podStartSLOduration=5.641244411 podStartE2EDuration="41.442712486s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:48.032554205 +0000 UTC m=+925.724162957" lastFinishedPulling="2025-12-03 06:43:23.83402228 +0000 UTC m=+961.525631032" observedRunningTime="2025-12-03 06:43:26.434678587 +0000 UTC m=+964.126287339" watchObservedRunningTime="2025-12-03 06:43:26.442712486 +0000 UTC m=+964.134321238" Dec 03 06:43:27 crc kubenswrapper[4818]: I1203 06:43:27.165196 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-98z98" event={"ID":"df292cb6-9f13-4cec-a7e1-1ef3155a7b1f","Type":"ContainerStarted","Data":"91f1e8cb18880749247d7d0a345e4e725879ea3ba61962992e70957caee9bd80"} Dec 03 06:43:27 crc kubenswrapper[4818]: I1203 06:43:27.165300 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-98z98" Dec 03 06:43:27 crc kubenswrapper[4818]: I1203 06:43:27.167165 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpnhz" event={"ID":"3d1565f0-62d4-4ecd-9d73-16338035196e","Type":"ContainerStarted","Data":"6c6285801cfdb950f5b6d11af4341213d4a8cfc84955109a77a6ef887e33f66e"} Dec 03 06:43:27 crc kubenswrapper[4818]: I1203 06:43:27.167262 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpnhz" Dec 03 06:43:27 crc kubenswrapper[4818]: I1203 06:43:27.169114 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wndt9" event={"ID":"91e645ae-28ec-4d3b-8fe3-e75aa1acf5a5","Type":"ContainerStarted","Data":"d7cbbdd87e1417a6b26350b6e8590906b50a54ef4bb51e8aead9270200ecda2c"} Dec 03 06:43:27 crc kubenswrapper[4818]: I1203 06:43:27.169228 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wndt9" Dec 03 06:43:27 crc kubenswrapper[4818]: I1203 06:43:27.171024 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bqx2t" event={"ID":"f4158ae0-a176-49a2-a8b8-57cd52efbdc3","Type":"ContainerStarted","Data":"4d1f833a8b92f3943daa52b76c493c7dd49112ccb8a7ac35e5cee1fd772130bc"} Dec 03 06:43:27 crc kubenswrapper[4818]: I1203 06:43:27.172532 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-4hdr6" event={"ID":"4130e7e0-4596-451f-b346-5afba08e9dea","Type":"ContainerStarted","Data":"6e439cd40bbd4fac4d23cdd90cf4c3c7c3b92f459be4a3578b8fc903882ac519"} Dec 03 06:43:27 crc kubenswrapper[4818]: I1203 06:43:27.199941 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-98z98" podStartSLOduration=4.299202944 podStartE2EDuration="42.199920983s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:47.742548654 +0000 UTC m=+925.434157406" lastFinishedPulling="2025-12-03 06:43:25.643266693 +0000 UTC m=+963.334875445" observedRunningTime="2025-12-03 06:43:27.193551815 +0000 UTC m=+964.885160567" watchObservedRunningTime="2025-12-03 06:43:27.199920983 +0000 UTC m=+964.891529735" Dec 03 06:43:27 crc kubenswrapper[4818]: I1203 06:43:27.213491 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wndt9" podStartSLOduration=3.657861731 podStartE2EDuration="42.213477249s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:46.993712735 +0000 UTC m=+924.685321487" lastFinishedPulling="2025-12-03 06:43:25.549328253 +0000 UTC m=+963.240937005" observedRunningTime="2025-12-03 06:43:27.209122551 +0000 UTC m=+964.900731303" watchObservedRunningTime="2025-12-03 06:43:27.213477249 +0000 UTC m=+964.905086001" Dec 03 06:43:27 crc kubenswrapper[4818]: I1203 06:43:27.228874 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpnhz" podStartSLOduration=4.034027208 podStartE2EDuration="42.22885447s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:47.454267345 +0000 UTC m=+925.145876097" lastFinishedPulling="2025-12-03 06:43:25.649094607 +0000 UTC m=+963.340703359" observedRunningTime="2025-12-03 06:43:27.226062221 +0000 UTC m=+964.917670973" watchObservedRunningTime="2025-12-03 06:43:27.22885447 +0000 UTC m=+964.920463222" Dec 03 06:43:27 crc kubenswrapper[4818]: I1203 06:43:27.261152 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bqx2t" podStartSLOduration=26.494660565 podStartE2EDuration="30.2611227s" podCreationTimestamp="2025-12-03 06:42:57 +0000 UTC" firstStartedPulling="2025-12-03 06:43:21.873729851 +0000 UTC m=+959.565338603" lastFinishedPulling="2025-12-03 06:43:25.640191986 +0000 UTC m=+963.331800738" observedRunningTime="2025-12-03 06:43:27.242755565 +0000 UTC m=+964.934364317" watchObservedRunningTime="2025-12-03 06:43:27.2611227 +0000 UTC m=+964.952731452" Dec 03 06:43:27 crc kubenswrapper[4818]: I1203 06:43:27.500212 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bqx2t" Dec 03 06:43:27 crc kubenswrapper[4818]: I1203 06:43:27.500273 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bqx2t" Dec 03 06:43:28 crc kubenswrapper[4818]: I1203 06:43:28.179289 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-4hdr6" Dec 03 06:43:28 crc kubenswrapper[4818]: I1203 06:43:28.753992 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-bqx2t" podUID="f4158ae0-a176-49a2-a8b8-57cd52efbdc3" containerName="registry-server" probeResult="failure" output=< Dec 03 06:43:28 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Dec 03 06:43:28 crc kubenswrapper[4818]: > Dec 03 06:43:31 crc kubenswrapper[4818]: I1203 06:43:31.658853 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gmkzd" Dec 03 06:43:31 crc kubenswrapper[4818]: I1203 06:43:31.685032 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-4hdr6" podStartSLOduration=7.957518317 podStartE2EDuration="46.685011948s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:46.958525972 +0000 UTC m=+924.650134724" lastFinishedPulling="2025-12-03 06:43:25.686019603 +0000 UTC m=+963.377628355" observedRunningTime="2025-12-03 06:43:27.267473818 +0000 UTC m=+964.959082570" watchObservedRunningTime="2025-12-03 06:43:31.685011948 +0000 UTC m=+969.376620710" Dec 03 06:43:32 crc kubenswrapper[4818]: I1203 06:43:32.101662 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd445pgf" Dec 03 06:43:32 crc kubenswrapper[4818]: I1203 06:43:32.632033 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-98d87bf44-8h8d6" Dec 03 06:43:35 crc kubenswrapper[4818]: I1203 06:43:35.615239 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-4hdr6" Dec 03 06:43:35 crc kubenswrapper[4818]: I1203 06:43:35.636745 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-wndt9" Dec 03 06:43:35 crc kubenswrapper[4818]: I1203 06:43:35.663173 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-2qvsd" Dec 03 06:43:35 crc kubenswrapper[4818]: I1203 06:43:35.742531 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-thzjt" Dec 03 06:43:35 crc kubenswrapper[4818]: I1203 06:43:35.845808 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpnhz" Dec 03 06:43:35 crc kubenswrapper[4818]: I1203 06:43:35.993914 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-98z98" Dec 03 06:43:36 crc kubenswrapper[4818]: I1203 06:43:36.363103 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xtq4s" Dec 03 06:43:37 crc kubenswrapper[4818]: I1203 06:43:37.538494 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bqx2t" Dec 03 06:43:37 crc kubenswrapper[4818]: I1203 06:43:37.577007 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bqx2t" Dec 03 06:43:37 crc kubenswrapper[4818]: I1203 06:43:37.774342 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bqx2t"] Dec 03 06:43:39 crc kubenswrapper[4818]: I1203 06:43:39.256443 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bqx2t" podUID="f4158ae0-a176-49a2-a8b8-57cd52efbdc3" containerName="registry-server" containerID="cri-o://4d1f833a8b92f3943daa52b76c493c7dd49112ccb8a7ac35e5cee1fd772130bc" gracePeriod=2 Dec 03 06:43:42 crc kubenswrapper[4818]: I1203 06:43:42.278333 4818 generic.go:334] "Generic (PLEG): container finished" podID="f4158ae0-a176-49a2-a8b8-57cd52efbdc3" containerID="4d1f833a8b92f3943daa52b76c493c7dd49112ccb8a7ac35e5cee1fd772130bc" exitCode=0 Dec 03 06:43:42 crc kubenswrapper[4818]: I1203 06:43:42.278425 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bqx2t" event={"ID":"f4158ae0-a176-49a2-a8b8-57cd52efbdc3","Type":"ContainerDied","Data":"4d1f833a8b92f3943daa52b76c493c7dd49112ccb8a7ac35e5cee1fd772130bc"} Dec 03 06:43:42 crc kubenswrapper[4818]: I1203 06:43:42.662828 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bqx2t" Dec 03 06:43:42 crc kubenswrapper[4818]: I1203 06:43:42.801348 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-catalog-content\") pod \"f4158ae0-a176-49a2-a8b8-57cd52efbdc3\" (UID: \"f4158ae0-a176-49a2-a8b8-57cd52efbdc3\") " Dec 03 06:43:42 crc kubenswrapper[4818]: I1203 06:43:42.801455 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcjpk\" (UniqueName: \"kubernetes.io/projected/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-kube-api-access-gcjpk\") pod \"f4158ae0-a176-49a2-a8b8-57cd52efbdc3\" (UID: \"f4158ae0-a176-49a2-a8b8-57cd52efbdc3\") " Dec 03 06:43:42 crc kubenswrapper[4818]: I1203 06:43:42.801538 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-utilities\") pod \"f4158ae0-a176-49a2-a8b8-57cd52efbdc3\" (UID: \"f4158ae0-a176-49a2-a8b8-57cd52efbdc3\") " Dec 03 06:43:42 crc kubenswrapper[4818]: I1203 06:43:42.802333 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-utilities" (OuterVolumeSpecName: "utilities") pod "f4158ae0-a176-49a2-a8b8-57cd52efbdc3" (UID: "f4158ae0-a176-49a2-a8b8-57cd52efbdc3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:43:42 crc kubenswrapper[4818]: I1203 06:43:42.806418 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-kube-api-access-gcjpk" (OuterVolumeSpecName: "kube-api-access-gcjpk") pod "f4158ae0-a176-49a2-a8b8-57cd52efbdc3" (UID: "f4158ae0-a176-49a2-a8b8-57cd52efbdc3"). InnerVolumeSpecName "kube-api-access-gcjpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:43:42 crc kubenswrapper[4818]: I1203 06:43:42.849449 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4158ae0-a176-49a2-a8b8-57cd52efbdc3" (UID: "f4158ae0-a176-49a2-a8b8-57cd52efbdc3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:43:42 crc kubenswrapper[4818]: I1203 06:43:42.903587 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:43:42 crc kubenswrapper[4818]: I1203 06:43:42.903624 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:43:42 crc kubenswrapper[4818]: I1203 06:43:42.903641 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcjpk\" (UniqueName: \"kubernetes.io/projected/f4158ae0-a176-49a2-a8b8-57cd52efbdc3-kube-api-access-gcjpk\") on node \"crc\" DevicePath \"\"" Dec 03 06:43:43 crc kubenswrapper[4818]: I1203 06:43:43.286475 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bqx2t" event={"ID":"f4158ae0-a176-49a2-a8b8-57cd52efbdc3","Type":"ContainerDied","Data":"7653fa9f81f80003eb880fe94a15686d2220bf1217884afa2d4f482990d0fa26"} Dec 03 06:43:43 crc kubenswrapper[4818]: I1203 06:43:43.287390 4818 scope.go:117] "RemoveContainer" containerID="4d1f833a8b92f3943daa52b76c493c7dd49112ccb8a7ac35e5cee1fd772130bc" Dec 03 06:43:43 crc kubenswrapper[4818]: I1203 06:43:43.287581 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bqx2t" Dec 03 06:43:43 crc kubenswrapper[4818]: I1203 06:43:43.302766 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:43:43 crc kubenswrapper[4818]: I1203 06:43:43.302880 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:43:43 crc kubenswrapper[4818]: I1203 06:43:43.309738 4818 scope.go:117] "RemoveContainer" containerID="af7698548d1e27d130f3deb9798775769cbde26be79f442ec2ca68667ca15b6a" Dec 03 06:43:43 crc kubenswrapper[4818]: I1203 06:43:43.333279 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bqx2t"] Dec 03 06:43:43 crc kubenswrapper[4818]: I1203 06:43:43.339514 4818 scope.go:117] "RemoveContainer" containerID="4c54f8a69a25fe4acc31175f5a5c6e830aaadb6ae44469cbdba5fa348eabf056" Dec 03 06:43:43 crc kubenswrapper[4818]: I1203 06:43:43.356453 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bqx2t"] Dec 03 06:43:44 crc kubenswrapper[4818]: I1203 06:43:44.748663 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4158ae0-a176-49a2-a8b8-57cd52efbdc3" path="/var/lib/kubelet/pods/f4158ae0-a176-49a2-a8b8-57cd52efbdc3/volumes" Dec 03 06:43:46 crc kubenswrapper[4818]: I1203 06:43:46.340125 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-27nrr" event={"ID":"1bd3fb88-7598-4a28-90fe-324d855dc90e","Type":"ContainerStarted","Data":"50ae4b9f704a547c9e224bf7227169a006beb54b4d5d5a68a2dd1e3f6e1b7977"} Dec 03 06:43:46 crc kubenswrapper[4818]: I1203 06:43:46.340718 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-27nrr" Dec 03 06:43:46 crc kubenswrapper[4818]: I1203 06:43:46.371621 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-27nrr" podStartSLOduration=3.156467948 podStartE2EDuration="1m1.371594456s" podCreationTimestamp="2025-12-03 06:42:45 +0000 UTC" firstStartedPulling="2025-12-03 06:42:47.463604997 +0000 UTC m=+925.155213749" lastFinishedPulling="2025-12-03 06:43:45.678731505 +0000 UTC m=+983.370340257" observedRunningTime="2025-12-03 06:43:46.36048583 +0000 UTC m=+984.052094612" watchObservedRunningTime="2025-12-03 06:43:46.371594456 +0000 UTC m=+984.063203228" Dec 03 06:43:55 crc kubenswrapper[4818]: I1203 06:43:55.867914 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-27nrr" Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.736320 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8b4gw"] Dec 03 06:44:11 crc kubenswrapper[4818]: E1203 06:44:11.737279 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4158ae0-a176-49a2-a8b8-57cd52efbdc3" containerName="extract-utilities" Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.737297 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4158ae0-a176-49a2-a8b8-57cd52efbdc3" containerName="extract-utilities" Dec 03 06:44:11 crc kubenswrapper[4818]: E1203 06:44:11.737339 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4158ae0-a176-49a2-a8b8-57cd52efbdc3" containerName="extract-content" Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.737347 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4158ae0-a176-49a2-a8b8-57cd52efbdc3" containerName="extract-content" Dec 03 06:44:11 crc kubenswrapper[4818]: E1203 06:44:11.737365 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4158ae0-a176-49a2-a8b8-57cd52efbdc3" containerName="registry-server" Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.737374 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4158ae0-a176-49a2-a8b8-57cd52efbdc3" containerName="registry-server" Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.737541 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4158ae0-a176-49a2-a8b8-57cd52efbdc3" containerName="registry-server" Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.740642 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-8b4gw" Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.742653 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.742785 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.742977 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-vp7g5" Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.743117 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.746673 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8b4gw"] Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.790298 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8sb8s"] Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.807774 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-8sb8s" Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.808532 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8sb8s"] Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.851555 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.929749 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e203487-144e-44d0-ba8e-f5374659484e-config\") pod \"dnsmasq-dns-78dd6ddcc-8sb8s\" (UID: \"1e203487-144e-44d0-ba8e-f5374659484e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8sb8s" Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.929973 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt8bn\" (UniqueName: \"kubernetes.io/projected/5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1-kube-api-access-pt8bn\") pod \"dnsmasq-dns-675f4bcbfc-8b4gw\" (UID: \"5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8b4gw" Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.930017 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1-config\") pod \"dnsmasq-dns-675f4bcbfc-8b4gw\" (UID: \"5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8b4gw" Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.930051 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e203487-144e-44d0-ba8e-f5374659484e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-8sb8s\" (UID: \"1e203487-144e-44d0-ba8e-f5374659484e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8sb8s" Dec 03 06:44:11 crc kubenswrapper[4818]: I1203 06:44:11.930295 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qx2m\" (UniqueName: \"kubernetes.io/projected/1e203487-144e-44d0-ba8e-f5374659484e-kube-api-access-9qx2m\") pod \"dnsmasq-dns-78dd6ddcc-8sb8s\" (UID: \"1e203487-144e-44d0-ba8e-f5374659484e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8sb8s" Dec 03 06:44:12 crc kubenswrapper[4818]: I1203 06:44:12.032136 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e203487-144e-44d0-ba8e-f5374659484e-config\") pod \"dnsmasq-dns-78dd6ddcc-8sb8s\" (UID: \"1e203487-144e-44d0-ba8e-f5374659484e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8sb8s" Dec 03 06:44:12 crc kubenswrapper[4818]: I1203 06:44:12.032221 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt8bn\" (UniqueName: \"kubernetes.io/projected/5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1-kube-api-access-pt8bn\") pod \"dnsmasq-dns-675f4bcbfc-8b4gw\" (UID: \"5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8b4gw" Dec 03 06:44:12 crc kubenswrapper[4818]: I1203 06:44:12.032254 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1-config\") pod \"dnsmasq-dns-675f4bcbfc-8b4gw\" (UID: \"5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8b4gw" Dec 03 06:44:12 crc kubenswrapper[4818]: I1203 06:44:12.032288 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e203487-144e-44d0-ba8e-f5374659484e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-8sb8s\" (UID: \"1e203487-144e-44d0-ba8e-f5374659484e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8sb8s" Dec 03 06:44:12 crc kubenswrapper[4818]: I1203 06:44:12.032353 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qx2m\" (UniqueName: \"kubernetes.io/projected/1e203487-144e-44d0-ba8e-f5374659484e-kube-api-access-9qx2m\") pod \"dnsmasq-dns-78dd6ddcc-8sb8s\" (UID: \"1e203487-144e-44d0-ba8e-f5374659484e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8sb8s" Dec 03 06:44:12 crc kubenswrapper[4818]: I1203 06:44:12.033803 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e203487-144e-44d0-ba8e-f5374659484e-config\") pod \"dnsmasq-dns-78dd6ddcc-8sb8s\" (UID: \"1e203487-144e-44d0-ba8e-f5374659484e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8sb8s" Dec 03 06:44:12 crc kubenswrapper[4818]: I1203 06:44:12.034744 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1-config\") pod \"dnsmasq-dns-675f4bcbfc-8b4gw\" (UID: \"5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8b4gw" Dec 03 06:44:12 crc kubenswrapper[4818]: I1203 06:44:12.035342 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e203487-144e-44d0-ba8e-f5374659484e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-8sb8s\" (UID: \"1e203487-144e-44d0-ba8e-f5374659484e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8sb8s" Dec 03 06:44:12 crc kubenswrapper[4818]: I1203 06:44:12.057491 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qx2m\" (UniqueName: \"kubernetes.io/projected/1e203487-144e-44d0-ba8e-f5374659484e-kube-api-access-9qx2m\") pod \"dnsmasq-dns-78dd6ddcc-8sb8s\" (UID: \"1e203487-144e-44d0-ba8e-f5374659484e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8sb8s" Dec 03 06:44:12 crc kubenswrapper[4818]: I1203 06:44:12.065943 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt8bn\" (UniqueName: \"kubernetes.io/projected/5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1-kube-api-access-pt8bn\") pod \"dnsmasq-dns-675f4bcbfc-8b4gw\" (UID: \"5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-8b4gw" Dec 03 06:44:12 crc kubenswrapper[4818]: I1203 06:44:12.162219 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-8sb8s" Dec 03 06:44:12 crc kubenswrapper[4818]: I1203 06:44:12.362990 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-8b4gw" Dec 03 06:44:12 crc kubenswrapper[4818]: I1203 06:44:12.584810 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8sb8s"] Dec 03 06:44:12 crc kubenswrapper[4818]: I1203 06:44:12.592243 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 06:44:12 crc kubenswrapper[4818]: I1203 06:44:12.776317 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8b4gw"] Dec 03 06:44:13 crc kubenswrapper[4818]: I1203 06:44:13.301858 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:44:13 crc kubenswrapper[4818]: I1203 06:44:13.301918 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:44:13 crc kubenswrapper[4818]: I1203 06:44:13.301969 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:44:13 crc kubenswrapper[4818]: I1203 06:44:13.302604 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9e30bff309bf8bc13ed470ed28234141d81432e279f1c0a4251c4793b659775c"} pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 06:44:13 crc kubenswrapper[4818]: I1203 06:44:13.302675 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" containerID="cri-o://9e30bff309bf8bc13ed470ed28234141d81432e279f1c0a4251c4793b659775c" gracePeriod=600 Dec 03 06:44:13 crc kubenswrapper[4818]: I1203 06:44:13.553747 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-8b4gw" event={"ID":"5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1","Type":"ContainerStarted","Data":"01bea1c39138b6ab82bd6e8e487a6b43af40e8e81a065aba4ae70db1785ebc1c"} Dec 03 06:44:13 crc kubenswrapper[4818]: I1203 06:44:13.556619 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-8sb8s" event={"ID":"1e203487-144e-44d0-ba8e-f5374659484e","Type":"ContainerStarted","Data":"74e5157a907a5cac1d21f7820bea92bc8afae3d1a9358592adf5964d8af5d585"} Dec 03 06:44:13 crc kubenswrapper[4818]: I1203 06:44:13.561731 4818 generic.go:334] "Generic (PLEG): container finished" podID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerID="9e30bff309bf8bc13ed470ed28234141d81432e279f1c0a4251c4793b659775c" exitCode=0 Dec 03 06:44:13 crc kubenswrapper[4818]: I1203 06:44:13.561773 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerDied","Data":"9e30bff309bf8bc13ed470ed28234141d81432e279f1c0a4251c4793b659775c"} Dec 03 06:44:13 crc kubenswrapper[4818]: I1203 06:44:13.561805 4818 scope.go:117] "RemoveContainer" containerID="d9ea0215376ae671f51a5593609153a4204668620744066ed3e3215741cd9e5f" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.260941 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8b4gw"] Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.302191 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-d25tm"] Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.303666 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-d25tm" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.320800 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-d25tm"] Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.377452 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b5ecd23-fa73-444e-a79c-166c4a25dff2-config\") pod \"dnsmasq-dns-666b6646f7-d25tm\" (UID: \"0b5ecd23-fa73-444e-a79c-166c4a25dff2\") " pod="openstack/dnsmasq-dns-666b6646f7-d25tm" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.377500 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b5ecd23-fa73-444e-a79c-166c4a25dff2-dns-svc\") pod \"dnsmasq-dns-666b6646f7-d25tm\" (UID: \"0b5ecd23-fa73-444e-a79c-166c4a25dff2\") " pod="openstack/dnsmasq-dns-666b6646f7-d25tm" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.377681 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk8ks\" (UniqueName: \"kubernetes.io/projected/0b5ecd23-fa73-444e-a79c-166c4a25dff2-kube-api-access-zk8ks\") pod \"dnsmasq-dns-666b6646f7-d25tm\" (UID: \"0b5ecd23-fa73-444e-a79c-166c4a25dff2\") " pod="openstack/dnsmasq-dns-666b6646f7-d25tm" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.478899 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk8ks\" (UniqueName: \"kubernetes.io/projected/0b5ecd23-fa73-444e-a79c-166c4a25dff2-kube-api-access-zk8ks\") pod \"dnsmasq-dns-666b6646f7-d25tm\" (UID: \"0b5ecd23-fa73-444e-a79c-166c4a25dff2\") " pod="openstack/dnsmasq-dns-666b6646f7-d25tm" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.478972 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b5ecd23-fa73-444e-a79c-166c4a25dff2-config\") pod \"dnsmasq-dns-666b6646f7-d25tm\" (UID: \"0b5ecd23-fa73-444e-a79c-166c4a25dff2\") " pod="openstack/dnsmasq-dns-666b6646f7-d25tm" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.478993 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b5ecd23-fa73-444e-a79c-166c4a25dff2-dns-svc\") pod \"dnsmasq-dns-666b6646f7-d25tm\" (UID: \"0b5ecd23-fa73-444e-a79c-166c4a25dff2\") " pod="openstack/dnsmasq-dns-666b6646f7-d25tm" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.480267 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b5ecd23-fa73-444e-a79c-166c4a25dff2-config\") pod \"dnsmasq-dns-666b6646f7-d25tm\" (UID: \"0b5ecd23-fa73-444e-a79c-166c4a25dff2\") " pod="openstack/dnsmasq-dns-666b6646f7-d25tm" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.480438 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b5ecd23-fa73-444e-a79c-166c4a25dff2-dns-svc\") pod \"dnsmasq-dns-666b6646f7-d25tm\" (UID: \"0b5ecd23-fa73-444e-a79c-166c4a25dff2\") " pod="openstack/dnsmasq-dns-666b6646f7-d25tm" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.607301 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8sb8s"] Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.609332 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"94f995293ac99f0a139bded43f6e45522bd84517d494216318f538517e1db600"} Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.629681 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk8ks\" (UniqueName: \"kubernetes.io/projected/0b5ecd23-fa73-444e-a79c-166c4a25dff2-kube-api-access-zk8ks\") pod \"dnsmasq-dns-666b6646f7-d25tm\" (UID: \"0b5ecd23-fa73-444e-a79c-166c4a25dff2\") " pod="openstack/dnsmasq-dns-666b6646f7-d25tm" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.637007 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-d25tm" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.643018 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f2vzn"] Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.644712 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-f2vzn" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.662483 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f2vzn"] Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.786419 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs9fk\" (UniqueName: \"kubernetes.io/projected/11a1627a-c749-4212-a52b-61e284a03211-kube-api-access-xs9fk\") pod \"dnsmasq-dns-57d769cc4f-f2vzn\" (UID: \"11a1627a-c749-4212-a52b-61e284a03211\") " pod="openstack/dnsmasq-dns-57d769cc4f-f2vzn" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.786474 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11a1627a-c749-4212-a52b-61e284a03211-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-f2vzn\" (UID: \"11a1627a-c749-4212-a52b-61e284a03211\") " pod="openstack/dnsmasq-dns-57d769cc4f-f2vzn" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.786500 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11a1627a-c749-4212-a52b-61e284a03211-config\") pod \"dnsmasq-dns-57d769cc4f-f2vzn\" (UID: \"11a1627a-c749-4212-a52b-61e284a03211\") " pod="openstack/dnsmasq-dns-57d769cc4f-f2vzn" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.966509 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs9fk\" (UniqueName: \"kubernetes.io/projected/11a1627a-c749-4212-a52b-61e284a03211-kube-api-access-xs9fk\") pod \"dnsmasq-dns-57d769cc4f-f2vzn\" (UID: \"11a1627a-c749-4212-a52b-61e284a03211\") " pod="openstack/dnsmasq-dns-57d769cc4f-f2vzn" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.966834 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11a1627a-c749-4212-a52b-61e284a03211-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-f2vzn\" (UID: \"11a1627a-c749-4212-a52b-61e284a03211\") " pod="openstack/dnsmasq-dns-57d769cc4f-f2vzn" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.966854 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11a1627a-c749-4212-a52b-61e284a03211-config\") pod \"dnsmasq-dns-57d769cc4f-f2vzn\" (UID: \"11a1627a-c749-4212-a52b-61e284a03211\") " pod="openstack/dnsmasq-dns-57d769cc4f-f2vzn" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.968667 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11a1627a-c749-4212-a52b-61e284a03211-config\") pod \"dnsmasq-dns-57d769cc4f-f2vzn\" (UID: \"11a1627a-c749-4212-a52b-61e284a03211\") " pod="openstack/dnsmasq-dns-57d769cc4f-f2vzn" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.968875 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11a1627a-c749-4212-a52b-61e284a03211-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-f2vzn\" (UID: \"11a1627a-c749-4212-a52b-61e284a03211\") " pod="openstack/dnsmasq-dns-57d769cc4f-f2vzn" Dec 03 06:44:14 crc kubenswrapper[4818]: I1203 06:44:14.984761 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs9fk\" (UniqueName: \"kubernetes.io/projected/11a1627a-c749-4212-a52b-61e284a03211-kube-api-access-xs9fk\") pod \"dnsmasq-dns-57d769cc4f-f2vzn\" (UID: \"11a1627a-c749-4212-a52b-61e284a03211\") " pod="openstack/dnsmasq-dns-57d769cc4f-f2vzn" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.142494 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-d25tm"] Dec 03 06:44:15 crc kubenswrapper[4818]: W1203 06:44:15.147420 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b5ecd23_fa73_444e_a79c_166c4a25dff2.slice/crio-c7bf2af9108106e72737e9c78d1d19076606c43f7efb1b68b8c57135c1f8154e WatchSource:0}: Error finding container c7bf2af9108106e72737e9c78d1d19076606c43f7efb1b68b8c57135c1f8154e: Status 404 returned error can't find the container with id c7bf2af9108106e72737e9c78d1d19076606c43f7efb1b68b8c57135c1f8154e Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.276276 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-f2vzn" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.520576 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.522581 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.525307 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.525483 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.525639 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.525743 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.525860 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-jws4q" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.531112 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.534533 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.540252 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.622474 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-d25tm" event={"ID":"0b5ecd23-fa73-444e-a79c-166c4a25dff2","Type":"ContainerStarted","Data":"c7bf2af9108106e72737e9c78d1d19076606c43f7efb1b68b8c57135c1f8154e"} Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.681278 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.681360 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.681391 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbbg7\" (UniqueName: \"kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-kube-api-access-dbbg7\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.681581 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.681628 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.681651 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.681686 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-config-data\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.681703 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.681721 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.681775 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0e751691-c660-4527-a9a6-61505f83c6ff-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.681809 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0e751691-c660-4527-a9a6-61505f83c6ff-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.783297 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.783373 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbbg7\" (UniqueName: \"kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-kube-api-access-dbbg7\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.783467 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.783488 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.783507 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.783533 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-config-data\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.783549 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.783563 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.783585 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0e751691-c660-4527-a9a6-61505f83c6ff-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.783620 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0e751691-c660-4527-a9a6-61505f83c6ff-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.784843 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.789618 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.790177 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.790416 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.790786 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.793251 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0e751691-c660-4527-a9a6-61505f83c6ff-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.793283 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.795547 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.797068 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.802849 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0e751691-c660-4527-a9a6-61505f83c6ff-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.803055 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-config-data\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.805160 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.806733 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.808839 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.809111 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.809140 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.809359 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.809494 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbbg7\" (UniqueName: \"kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-kube-api-access-dbbg7\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.812293 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.813382 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-pnlrk" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.813606 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.814340 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.841261 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.865872 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.877858 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f2vzn"] Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.988883 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.988941 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.988960 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.988982 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.989010 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df9ab7c7-152b-4938-9ef5-d0588111b8a2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.989049 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df9ab7c7-152b-4938-9ef5-d0588111b8a2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.989240 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.989286 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.989320 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.989355 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:15 crc kubenswrapper[4818]: I1203 06:44:15.989393 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pwn2\" (UniqueName: \"kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-kube-api-access-8pwn2\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.093809 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.093892 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df9ab7c7-152b-4938-9ef5-d0588111b8a2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.093939 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df9ab7c7-152b-4938-9ef5-d0588111b8a2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.094002 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.094036 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.094067 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.094101 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.094137 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pwn2\" (UniqueName: \"kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-kube-api-access-8pwn2\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.094160 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.094186 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.094208 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.095207 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.095431 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.097000 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.100126 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.100556 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.100862 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.118831 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df9ab7c7-152b-4938-9ef5-d0588111b8a2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.184152 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.184317 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.184711 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df9ab7c7-152b-4938-9ef5-d0588111b8a2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.188039 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.189254 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pwn2\" (UniqueName: \"kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-kube-api-access-8pwn2\") pod \"rabbitmq-cell1-server-0\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.255323 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.635690 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f2vzn" event={"ID":"11a1627a-c749-4212-a52b-61e284a03211","Type":"ContainerStarted","Data":"d1ca1bcdc4619647d1bb88cb5dfa3fa5e04781cb08993f2e994add6d3a35af35"} Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.686259 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 06:44:16 crc kubenswrapper[4818]: W1203 06:44:16.696108 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e751691_c660_4527_a9a6_61505f83c6ff.slice/crio-7ee3814f8164ad501c7750b02416b4c23efac5aded2434000a581b60871030d6 WatchSource:0}: Error finding container 7ee3814f8164ad501c7750b02416b4c23efac5aded2434000a581b60871030d6: Status 404 returned error can't find the container with id 7ee3814f8164ad501c7750b02416b4c23efac5aded2434000a581b60871030d6 Dec 03 06:44:16 crc kubenswrapper[4818]: I1203 06:44:16.827770 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 06:44:16 crc kubenswrapper[4818]: W1203 06:44:16.849422 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf9ab7c7_152b_4938_9ef5_d0588111b8a2.slice/crio-13eb95568ccbe82db9fe67086da897b8d23545d092808a160162c9904584eeda WatchSource:0}: Error finding container 13eb95568ccbe82db9fe67086da897b8d23545d092808a160162c9904584eeda: Status 404 returned error can't find the container with id 13eb95568ccbe82db9fe67086da897b8d23545d092808a160162c9904584eeda Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.063089 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.064924 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.069367 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.070314 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-j77wq" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.070722 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.073243 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.073936 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.096097 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.242110 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/206b044a-6264-4c4e-89b9-7cacc4f30ac5-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.242243 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/206b044a-6264-4c4e-89b9-7cacc4f30ac5-operator-scripts\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.242286 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.242321 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkrbz\" (UniqueName: \"kubernetes.io/projected/206b044a-6264-4c4e-89b9-7cacc4f30ac5-kube-api-access-zkrbz\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.242366 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/206b044a-6264-4c4e-89b9-7cacc4f30ac5-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.242404 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/206b044a-6264-4c4e-89b9-7cacc4f30ac5-kolla-config\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.242448 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/206b044a-6264-4c4e-89b9-7cacc4f30ac5-config-data-default\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.243501 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/206b044a-6264-4c4e-89b9-7cacc4f30ac5-config-data-generated\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.346619 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/206b044a-6264-4c4e-89b9-7cacc4f30ac5-operator-scripts\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.346671 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.346699 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkrbz\" (UniqueName: \"kubernetes.io/projected/206b044a-6264-4c4e-89b9-7cacc4f30ac5-kube-api-access-zkrbz\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.346719 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/206b044a-6264-4c4e-89b9-7cacc4f30ac5-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.346746 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/206b044a-6264-4c4e-89b9-7cacc4f30ac5-kolla-config\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.346776 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/206b044a-6264-4c4e-89b9-7cacc4f30ac5-config-data-default\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.346798 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/206b044a-6264-4c4e-89b9-7cacc4f30ac5-config-data-generated\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.346852 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/206b044a-6264-4c4e-89b9-7cacc4f30ac5-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.348054 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.349723 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/206b044a-6264-4c4e-89b9-7cacc4f30ac5-config-data-default\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.350041 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/206b044a-6264-4c4e-89b9-7cacc4f30ac5-config-data-generated\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.350214 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/206b044a-6264-4c4e-89b9-7cacc4f30ac5-kolla-config\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.351655 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/206b044a-6264-4c4e-89b9-7cacc4f30ac5-operator-scripts\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.359452 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/206b044a-6264-4c4e-89b9-7cacc4f30ac5-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.366611 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/206b044a-6264-4c4e-89b9-7cacc4f30ac5-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.370316 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkrbz\" (UniqueName: \"kubernetes.io/projected/206b044a-6264-4c4e-89b9-7cacc4f30ac5-kube-api-access-zkrbz\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.388058 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"206b044a-6264-4c4e-89b9-7cacc4f30ac5\") " pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.396583 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.707780 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"df9ab7c7-152b-4938-9ef5-d0588111b8a2","Type":"ContainerStarted","Data":"13eb95568ccbe82db9fe67086da897b8d23545d092808a160162c9904584eeda"} Dec 03 06:44:17 crc kubenswrapper[4818]: I1203 06:44:17.719929 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0e751691-c660-4527-a9a6-61505f83c6ff","Type":"ContainerStarted","Data":"7ee3814f8164ad501c7750b02416b4c23efac5aded2434000a581b60871030d6"} Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.037689 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.783685 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.785442 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.785677 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.789637 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-jp8gq" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.790015 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.790153 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.790262 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.895652 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dd3e13-e5ee-4315-ab7a-222c902192be-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.895765 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0dd3e13-e5ee-4315-ab7a-222c902192be-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.895850 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a0dd3e13-e5ee-4315-ab7a-222c902192be-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.895879 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0dd3e13-e5ee-4315-ab7a-222c902192be-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.895934 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.895970 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a0dd3e13-e5ee-4315-ab7a-222c902192be-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.895994 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a0dd3e13-e5ee-4315-ab7a-222c902192be-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.896042 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89j2d\" (UniqueName: \"kubernetes.io/projected/a0dd3e13-e5ee-4315-ab7a-222c902192be-kube-api-access-89j2d\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.905949 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.906869 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.912843 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.913011 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-p5fvf" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.913190 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.917577 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.998311 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dd3e13-e5ee-4315-ab7a-222c902192be-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.998377 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0dd3e13-e5ee-4315-ab7a-222c902192be-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.998417 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a0dd3e13-e5ee-4315-ab7a-222c902192be-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.998443 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0dd3e13-e5ee-4315-ab7a-222c902192be-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.998482 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.998503 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a0dd3e13-e5ee-4315-ab7a-222c902192be-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.998517 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a0dd3e13-e5ee-4315-ab7a-222c902192be-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.998556 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89j2d\" (UniqueName: \"kubernetes.io/projected/a0dd3e13-e5ee-4315-ab7a-222c902192be-kube-api-access-89j2d\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:18 crc kubenswrapper[4818]: I1203 06:44:18.999203 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.007860 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a0dd3e13-e5ee-4315-ab7a-222c902192be-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.008284 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a0dd3e13-e5ee-4315-ab7a-222c902192be-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.008472 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a0dd3e13-e5ee-4315-ab7a-222c902192be-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.011647 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0dd3e13-e5ee-4315-ab7a-222c902192be-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.017400 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0dd3e13-e5ee-4315-ab7a-222c902192be-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.020269 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0dd3e13-e5ee-4315-ab7a-222c902192be-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.022126 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89j2d\" (UniqueName: \"kubernetes.io/projected/a0dd3e13-e5ee-4315-ab7a-222c902192be-kube-api-access-89j2d\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.034837 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a0dd3e13-e5ee-4315-ab7a-222c902192be\") " pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.100493 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c826858d-45cb-4ae3-8fa2-1531f9390480-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c826858d-45cb-4ae3-8fa2-1531f9390480\") " pod="openstack/memcached-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.100545 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9t9s\" (UniqueName: \"kubernetes.io/projected/c826858d-45cb-4ae3-8fa2-1531f9390480-kube-api-access-c9t9s\") pod \"memcached-0\" (UID: \"c826858d-45cb-4ae3-8fa2-1531f9390480\") " pod="openstack/memcached-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.100571 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c826858d-45cb-4ae3-8fa2-1531f9390480-config-data\") pod \"memcached-0\" (UID: \"c826858d-45cb-4ae3-8fa2-1531f9390480\") " pod="openstack/memcached-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.100592 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c826858d-45cb-4ae3-8fa2-1531f9390480-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c826858d-45cb-4ae3-8fa2-1531f9390480\") " pod="openstack/memcached-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.100697 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c826858d-45cb-4ae3-8fa2-1531f9390480-kolla-config\") pod \"memcached-0\" (UID: \"c826858d-45cb-4ae3-8fa2-1531f9390480\") " pod="openstack/memcached-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.130601 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.202033 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c826858d-45cb-4ae3-8fa2-1531f9390480-kolla-config\") pod \"memcached-0\" (UID: \"c826858d-45cb-4ae3-8fa2-1531f9390480\") " pod="openstack/memcached-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.202183 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c826858d-45cb-4ae3-8fa2-1531f9390480-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c826858d-45cb-4ae3-8fa2-1531f9390480\") " pod="openstack/memcached-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.202211 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9t9s\" (UniqueName: \"kubernetes.io/projected/c826858d-45cb-4ae3-8fa2-1531f9390480-kube-api-access-c9t9s\") pod \"memcached-0\" (UID: \"c826858d-45cb-4ae3-8fa2-1531f9390480\") " pod="openstack/memcached-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.202256 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c826858d-45cb-4ae3-8fa2-1531f9390480-config-data\") pod \"memcached-0\" (UID: \"c826858d-45cb-4ae3-8fa2-1531f9390480\") " pod="openstack/memcached-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.202280 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c826858d-45cb-4ae3-8fa2-1531f9390480-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c826858d-45cb-4ae3-8fa2-1531f9390480\") " pod="openstack/memcached-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.203292 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c826858d-45cb-4ae3-8fa2-1531f9390480-kolla-config\") pod \"memcached-0\" (UID: \"c826858d-45cb-4ae3-8fa2-1531f9390480\") " pod="openstack/memcached-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.203504 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c826858d-45cb-4ae3-8fa2-1531f9390480-config-data\") pod \"memcached-0\" (UID: \"c826858d-45cb-4ae3-8fa2-1531f9390480\") " pod="openstack/memcached-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.206328 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c826858d-45cb-4ae3-8fa2-1531f9390480-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c826858d-45cb-4ae3-8fa2-1531f9390480\") " pod="openstack/memcached-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.206443 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c826858d-45cb-4ae3-8fa2-1531f9390480-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c826858d-45cb-4ae3-8fa2-1531f9390480\") " pod="openstack/memcached-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.222710 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9t9s\" (UniqueName: \"kubernetes.io/projected/c826858d-45cb-4ae3-8fa2-1531f9390480-kube-api-access-c9t9s\") pod \"memcached-0\" (UID: \"c826858d-45cb-4ae3-8fa2-1531f9390480\") " pod="openstack/memcached-0" Dec 03 06:44:19 crc kubenswrapper[4818]: I1203 06:44:19.238986 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 06:44:21 crc kubenswrapper[4818]: I1203 06:44:21.057759 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 06:44:21 crc kubenswrapper[4818]: I1203 06:44:21.059153 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 06:44:21 crc kubenswrapper[4818]: I1203 06:44:21.063722 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-d4sw7" Dec 03 06:44:21 crc kubenswrapper[4818]: I1203 06:44:21.067415 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 06:44:21 crc kubenswrapper[4818]: I1203 06:44:21.151896 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl8tq\" (UniqueName: \"kubernetes.io/projected/644cd011-5c7b-4050-85e1-2194f5cd8d93-kube-api-access-pl8tq\") pod \"kube-state-metrics-0\" (UID: \"644cd011-5c7b-4050-85e1-2194f5cd8d93\") " pod="openstack/kube-state-metrics-0" Dec 03 06:44:21 crc kubenswrapper[4818]: I1203 06:44:21.253047 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl8tq\" (UniqueName: \"kubernetes.io/projected/644cd011-5c7b-4050-85e1-2194f5cd8d93-kube-api-access-pl8tq\") pod \"kube-state-metrics-0\" (UID: \"644cd011-5c7b-4050-85e1-2194f5cd8d93\") " pod="openstack/kube-state-metrics-0" Dec 03 06:44:21 crc kubenswrapper[4818]: I1203 06:44:21.275848 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl8tq\" (UniqueName: \"kubernetes.io/projected/644cd011-5c7b-4050-85e1-2194f5cd8d93-kube-api-access-pl8tq\") pod \"kube-state-metrics-0\" (UID: \"644cd011-5c7b-4050-85e1-2194f5cd8d93\") " pod="openstack/kube-state-metrics-0" Dec 03 06:44:21 crc kubenswrapper[4818]: I1203 06:44:21.388866 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 06:44:24 crc kubenswrapper[4818]: I1203 06:44:24.807606 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"206b044a-6264-4c4e-89b9-7cacc4f30ac5","Type":"ContainerStarted","Data":"48aa5ea37a4926ce6e349d242bda163046c79f76e274ee585aae2f8972155641"} Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.033607 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-kvg5t"] Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.034771 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.043263 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.043538 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.043660 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-5p8nd" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.051414 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kvg5t"] Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.061896 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-brxh4"] Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.063992 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.074766 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-brxh4"] Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.234964 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b5f31813-ba17-4452-9b1a-ee7dcfac564e-var-run\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.235024 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04bf523a-de6a-49a0-93ad-cca1e22939c6-scripts\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.235061 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04bf523a-de6a-49a0-93ad-cca1e22939c6-combined-ca-bundle\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.235100 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzhpk\" (UniqueName: \"kubernetes.io/projected/04bf523a-de6a-49a0-93ad-cca1e22939c6-kube-api-access-zzhpk\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.235147 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/04bf523a-de6a-49a0-93ad-cca1e22939c6-ovn-controller-tls-certs\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.235174 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjgjz\" (UniqueName: \"kubernetes.io/projected/b5f31813-ba17-4452-9b1a-ee7dcfac564e-kube-api-access-gjgjz\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.235195 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5f31813-ba17-4452-9b1a-ee7dcfac564e-scripts\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.235218 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b5f31813-ba17-4452-9b1a-ee7dcfac564e-var-lib\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.235246 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b5f31813-ba17-4452-9b1a-ee7dcfac564e-var-log\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.235284 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/04bf523a-de6a-49a0-93ad-cca1e22939c6-var-run-ovn\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.235325 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b5f31813-ba17-4452-9b1a-ee7dcfac564e-etc-ovs\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.235942 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/04bf523a-de6a-49a0-93ad-cca1e22939c6-var-run\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.235994 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/04bf523a-de6a-49a0-93ad-cca1e22939c6-var-log-ovn\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337016 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzhpk\" (UniqueName: \"kubernetes.io/projected/04bf523a-de6a-49a0-93ad-cca1e22939c6-kube-api-access-zzhpk\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337093 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/04bf523a-de6a-49a0-93ad-cca1e22939c6-ovn-controller-tls-certs\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337119 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjgjz\" (UniqueName: \"kubernetes.io/projected/b5f31813-ba17-4452-9b1a-ee7dcfac564e-kube-api-access-gjgjz\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337136 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5f31813-ba17-4452-9b1a-ee7dcfac564e-scripts\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337155 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b5f31813-ba17-4452-9b1a-ee7dcfac564e-var-lib\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337176 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b5f31813-ba17-4452-9b1a-ee7dcfac564e-var-log\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337205 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/04bf523a-de6a-49a0-93ad-cca1e22939c6-var-run-ovn\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337221 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b5f31813-ba17-4452-9b1a-ee7dcfac564e-etc-ovs\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337246 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/04bf523a-de6a-49a0-93ad-cca1e22939c6-var-run\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337271 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/04bf523a-de6a-49a0-93ad-cca1e22939c6-var-log-ovn\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337300 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b5f31813-ba17-4452-9b1a-ee7dcfac564e-var-run\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337322 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04bf523a-de6a-49a0-93ad-cca1e22939c6-scripts\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337346 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04bf523a-de6a-49a0-93ad-cca1e22939c6-combined-ca-bundle\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337645 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b5f31813-ba17-4452-9b1a-ee7dcfac564e-etc-ovs\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337756 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b5f31813-ba17-4452-9b1a-ee7dcfac564e-var-log\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337855 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b5f31813-ba17-4452-9b1a-ee7dcfac564e-var-run\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337929 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/04bf523a-de6a-49a0-93ad-cca1e22939c6-var-run-ovn\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.337935 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/04bf523a-de6a-49a0-93ad-cca1e22939c6-var-run\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.338029 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/04bf523a-de6a-49a0-93ad-cca1e22939c6-var-log-ovn\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.338087 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b5f31813-ba17-4452-9b1a-ee7dcfac564e-var-lib\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.339978 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5f31813-ba17-4452-9b1a-ee7dcfac564e-scripts\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.340690 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04bf523a-de6a-49a0-93ad-cca1e22939c6-scripts\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.351057 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/04bf523a-de6a-49a0-93ad-cca1e22939c6-ovn-controller-tls-certs\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.351334 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04bf523a-de6a-49a0-93ad-cca1e22939c6-combined-ca-bundle\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.356278 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzhpk\" (UniqueName: \"kubernetes.io/projected/04bf523a-de6a-49a0-93ad-cca1e22939c6-kube-api-access-zzhpk\") pod \"ovn-controller-kvg5t\" (UID: \"04bf523a-de6a-49a0-93ad-cca1e22939c6\") " pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.356426 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjgjz\" (UniqueName: \"kubernetes.io/projected/b5f31813-ba17-4452-9b1a-ee7dcfac564e-kube-api-access-gjgjz\") pod \"ovn-controller-ovs-brxh4\" (UID: \"b5f31813-ba17-4452-9b1a-ee7dcfac564e\") " pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.416014 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:25 crc kubenswrapper[4818]: I1203 06:44:25.440332 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.010555 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.012604 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.025795 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-g6xjc" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.027731 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.027882 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.027932 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.031727 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.062629 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.209047 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b27dc29-202f-4efb-9545-031eaa1d7da3-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.209095 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6xm5\" (UniqueName: \"kubernetes.io/projected/9b27dc29-202f-4efb-9545-031eaa1d7da3-kube-api-access-r6xm5\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.209120 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b27dc29-202f-4efb-9545-031eaa1d7da3-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.209137 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b27dc29-202f-4efb-9545-031eaa1d7da3-config\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.209178 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b27dc29-202f-4efb-9545-031eaa1d7da3-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.209195 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.209376 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9b27dc29-202f-4efb-9545-031eaa1d7da3-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.209406 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b27dc29-202f-4efb-9545-031eaa1d7da3-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.311294 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b27dc29-202f-4efb-9545-031eaa1d7da3-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.311359 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6xm5\" (UniqueName: \"kubernetes.io/projected/9b27dc29-202f-4efb-9545-031eaa1d7da3-kube-api-access-r6xm5\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.311381 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b27dc29-202f-4efb-9545-031eaa1d7da3-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.311398 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b27dc29-202f-4efb-9545-031eaa1d7da3-config\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.311440 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b27dc29-202f-4efb-9545-031eaa1d7da3-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.311460 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.311484 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9b27dc29-202f-4efb-9545-031eaa1d7da3-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.311510 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b27dc29-202f-4efb-9545-031eaa1d7da3-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.313423 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b27dc29-202f-4efb-9545-031eaa1d7da3-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.313727 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9b27dc29-202f-4efb-9545-031eaa1d7da3-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.313787 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b27dc29-202f-4efb-9545-031eaa1d7da3-config\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.313795 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.316193 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b27dc29-202f-4efb-9545-031eaa1d7da3-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.317702 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b27dc29-202f-4efb-9545-031eaa1d7da3-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.322684 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b27dc29-202f-4efb-9545-031eaa1d7da3-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.327050 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6xm5\" (UniqueName: \"kubernetes.io/projected/9b27dc29-202f-4efb-9545-031eaa1d7da3-kube-api-access-r6xm5\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.365546 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9b27dc29-202f-4efb-9545-031eaa1d7da3\") " pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:26 crc kubenswrapper[4818]: I1203 06:44:26.387646 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:28 crc kubenswrapper[4818]: I1203 06:44:28.884343 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 06:44:28 crc kubenswrapper[4818]: I1203 06:44:28.887009 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:28 crc kubenswrapper[4818]: I1203 06:44:28.889566 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 03 06:44:28 crc kubenswrapper[4818]: I1203 06:44:28.890593 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 03 06:44:28 crc kubenswrapper[4818]: I1203 06:44:28.890683 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 03 06:44:28 crc kubenswrapper[4818]: I1203 06:44:28.891598 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-zdtmj" Dec 03 06:44:28 crc kubenswrapper[4818]: I1203 06:44:28.901993 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.058427 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a27ef44f-2bff-4714-8018-a50f2f2de24a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.058973 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a27ef44f-2bff-4714-8018-a50f2f2de24a-config\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.059168 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a27ef44f-2bff-4714-8018-a50f2f2de24a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.059325 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a27ef44f-2bff-4714-8018-a50f2f2de24a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.059488 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a27ef44f-2bff-4714-8018-a50f2f2de24a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.059631 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75wdf\" (UniqueName: \"kubernetes.io/projected/a27ef44f-2bff-4714-8018-a50f2f2de24a-kube-api-access-75wdf\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.060102 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a27ef44f-2bff-4714-8018-a50f2f2de24a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.060255 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.161733 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.161805 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a27ef44f-2bff-4714-8018-a50f2f2de24a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.161848 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a27ef44f-2bff-4714-8018-a50f2f2de24a-config\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.161887 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a27ef44f-2bff-4714-8018-a50f2f2de24a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.161907 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a27ef44f-2bff-4714-8018-a50f2f2de24a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.161947 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a27ef44f-2bff-4714-8018-a50f2f2de24a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.161995 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75wdf\" (UniqueName: \"kubernetes.io/projected/a27ef44f-2bff-4714-8018-a50f2f2de24a-kube-api-access-75wdf\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.162064 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a27ef44f-2bff-4714-8018-a50f2f2de24a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.162079 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.162959 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a27ef44f-2bff-4714-8018-a50f2f2de24a-config\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.163080 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a27ef44f-2bff-4714-8018-a50f2f2de24a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.163222 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a27ef44f-2bff-4714-8018-a50f2f2de24a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.168158 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a27ef44f-2bff-4714-8018-a50f2f2de24a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.168214 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a27ef44f-2bff-4714-8018-a50f2f2de24a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.182097 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a27ef44f-2bff-4714-8018-a50f2f2de24a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.184877 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75wdf\" (UniqueName: \"kubernetes.io/projected/a27ef44f-2bff-4714-8018-a50f2f2de24a-kube-api-access-75wdf\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.188675 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"a27ef44f-2bff-4714-8018-a50f2f2de24a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:29 crc kubenswrapper[4818]: I1203 06:44:29.220655 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:30 crc kubenswrapper[4818]: I1203 06:44:30.906306 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 06:44:31 crc kubenswrapper[4818]: E1203 06:44:31.897613 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 03 06:44:31 crc kubenswrapper[4818]: E1203 06:44:31.898154 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8pwn2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(df9ab7c7-152b-4938-9ef5-d0588111b8a2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:44:31 crc kubenswrapper[4818]: E1203 06:44:31.899373 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="df9ab7c7-152b-4938-9ef5-d0588111b8a2" Dec 03 06:44:32 crc kubenswrapper[4818]: E1203 06:44:32.874238 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="df9ab7c7-152b-4938-9ef5-d0588111b8a2" Dec 03 06:44:36 crc kubenswrapper[4818]: W1203 06:44:36.166910 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0dd3e13_e5ee_4315_ab7a_222c902192be.slice/crio-8346d07b74d27e40efc7daa16c83cc8aa7e525869ce62bae30c57614c02be712 WatchSource:0}: Error finding container 8346d07b74d27e40efc7daa16c83cc8aa7e525869ce62bae30c57614c02be712: Status 404 returned error can't find the container with id 8346d07b74d27e40efc7daa16c83cc8aa7e525869ce62bae30c57614c02be712 Dec 03 06:44:36 crc kubenswrapper[4818]: E1203 06:44:36.167994 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 03 06:44:36 crc kubenswrapper[4818]: E1203 06:44:36.168150 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dbbg7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(0e751691-c660-4527-a9a6-61505f83c6ff): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:44:36 crc kubenswrapper[4818]: E1203 06:44:36.169305 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="0e751691-c660-4527-a9a6-61505f83c6ff" Dec 03 06:44:36 crc kubenswrapper[4818]: I1203 06:44:36.899082 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a0dd3e13-e5ee-4315-ab7a-222c902192be","Type":"ContainerStarted","Data":"8346d07b74d27e40efc7daa16c83cc8aa7e525869ce62bae30c57614c02be712"} Dec 03 06:44:36 crc kubenswrapper[4818]: E1203 06:44:36.901048 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="0e751691-c660-4527-a9a6-61505f83c6ff" Dec 03 06:44:37 crc kubenswrapper[4818]: E1203 06:44:37.044319 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 06:44:37 crc kubenswrapper[4818]: E1203 06:44:37.044888 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zk8ks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-d25tm_openstack(0b5ecd23-fa73-444e-a79c-166c4a25dff2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:44:37 crc kubenswrapper[4818]: E1203 06:44:37.046312 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-d25tm" podUID="0b5ecd23-fa73-444e-a79c-166c4a25dff2" Dec 03 06:44:37 crc kubenswrapper[4818]: E1203 06:44:37.103711 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 06:44:37 crc kubenswrapper[4818]: E1203 06:44:37.103865 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pt8bn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-8b4gw_openstack(5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:44:37 crc kubenswrapper[4818]: E1203 06:44:37.105287 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-8b4gw" podUID="5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1" Dec 03 06:44:37 crc kubenswrapper[4818]: E1203 06:44:37.116776 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 06:44:37 crc kubenswrapper[4818]: E1203 06:44:37.120029 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9qx2m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-8sb8s_openstack(1e203487-144e-44d0-ba8e-f5374659484e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:44:37 crc kubenswrapper[4818]: E1203 06:44:37.121960 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-8sb8s" podUID="1e203487-144e-44d0-ba8e-f5374659484e" Dec 03 06:44:37 crc kubenswrapper[4818]: E1203 06:44:37.131031 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 06:44:37 crc kubenswrapper[4818]: E1203 06:44:37.131185 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xs9fk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-f2vzn_openstack(11a1627a-c749-4212-a52b-61e284a03211): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:44:37 crc kubenswrapper[4818]: E1203 06:44:37.132326 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-f2vzn" podUID="11a1627a-c749-4212-a52b-61e284a03211" Dec 03 06:44:37 crc kubenswrapper[4818]: I1203 06:44:37.450205 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 06:44:37 crc kubenswrapper[4818]: I1203 06:44:37.504388 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 06:44:37 crc kubenswrapper[4818]: I1203 06:44:37.634517 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-brxh4"] Dec 03 06:44:37 crc kubenswrapper[4818]: E1203 06:44:37.908773 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-f2vzn" podUID="11a1627a-c749-4212-a52b-61e284a03211" Dec 03 06:44:37 crc kubenswrapper[4818]: E1203 06:44:37.908871 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-d25tm" podUID="0b5ecd23-fa73-444e-a79c-166c4a25dff2" Dec 03 06:44:38 crc kubenswrapper[4818]: I1203 06:44:38.557141 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 06:44:38 crc kubenswrapper[4818]: W1203 06:44:38.948398 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5f31813_ba17_4452_9b1a_ee7dcfac564e.slice/crio-6261c85436b24d45d624d8d0854e046566e5fc30778b276cf9f3d993ffef1cbb WatchSource:0}: Error finding container 6261c85436b24d45d624d8d0854e046566e5fc30778b276cf9f3d993ffef1cbb: Status 404 returned error can't find the container with id 6261c85436b24d45d624d8d0854e046566e5fc30778b276cf9f3d993ffef1cbb Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.084309 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-8sb8s" Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.105234 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-8b4gw" Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.111442 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e203487-144e-44d0-ba8e-f5374659484e-config\") pod \"1e203487-144e-44d0-ba8e-f5374659484e\" (UID: \"1e203487-144e-44d0-ba8e-f5374659484e\") " Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.111537 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e203487-144e-44d0-ba8e-f5374659484e-dns-svc\") pod \"1e203487-144e-44d0-ba8e-f5374659484e\" (UID: \"1e203487-144e-44d0-ba8e-f5374659484e\") " Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.111671 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qx2m\" (UniqueName: \"kubernetes.io/projected/1e203487-144e-44d0-ba8e-f5374659484e-kube-api-access-9qx2m\") pod \"1e203487-144e-44d0-ba8e-f5374659484e\" (UID: \"1e203487-144e-44d0-ba8e-f5374659484e\") " Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.114319 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e203487-144e-44d0-ba8e-f5374659484e-config" (OuterVolumeSpecName: "config") pod "1e203487-144e-44d0-ba8e-f5374659484e" (UID: "1e203487-144e-44d0-ba8e-f5374659484e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.114890 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e203487-144e-44d0-ba8e-f5374659484e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1e203487-144e-44d0-ba8e-f5374659484e" (UID: "1e203487-144e-44d0-ba8e-f5374659484e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.118021 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e203487-144e-44d0-ba8e-f5374659484e-kube-api-access-9qx2m" (OuterVolumeSpecName: "kube-api-access-9qx2m") pod "1e203487-144e-44d0-ba8e-f5374659484e" (UID: "1e203487-144e-44d0-ba8e-f5374659484e"). InnerVolumeSpecName "kube-api-access-9qx2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.213991 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pt8bn\" (UniqueName: \"kubernetes.io/projected/5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1-kube-api-access-pt8bn\") pod \"5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1\" (UID: \"5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1\") " Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.214327 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1-config\") pod \"5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1\" (UID: \"5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1\") " Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.214708 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1-config" (OuterVolumeSpecName: "config") pod "5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1" (UID: "5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.214798 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e203487-144e-44d0-ba8e-f5374659484e-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.214834 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e203487-144e-44d0-ba8e-f5374659484e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.214849 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qx2m\" (UniqueName: \"kubernetes.io/projected/1e203487-144e-44d0-ba8e-f5374659484e-kube-api-access-9qx2m\") on node \"crc\" DevicePath \"\"" Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.218628 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1-kube-api-access-pt8bn" (OuterVolumeSpecName: "kube-api-access-pt8bn") pod "5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1" (UID: "5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1"). InnerVolumeSpecName "kube-api-access-pt8bn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.315964 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pt8bn\" (UniqueName: \"kubernetes.io/projected/5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1-kube-api-access-pt8bn\") on node \"crc\" DevicePath \"\"" Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.315991 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.455993 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kvg5t"] Dec 03 06:44:39 crc kubenswrapper[4818]: W1203 06:44:39.461243 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04bf523a_de6a_49a0_93ad_cca1e22939c6.slice/crio-55b1e5eab4d67166789ef3f3e15203da4b546022c2852f9906d24167245ec410 WatchSource:0}: Error finding container 55b1e5eab4d67166789ef3f3e15203da4b546022c2852f9906d24167245ec410: Status 404 returned error can't find the container with id 55b1e5eab4d67166789ef3f3e15203da4b546022c2852f9906d24167245ec410 Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.708189 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 06:44:39 crc kubenswrapper[4818]: W1203 06:44:39.893917 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda27ef44f_2bff_4714_8018_a50f2f2de24a.slice/crio-55c7cb29ed4044abb3616655f924ef240e0a85e0510f904d15e60a9677d9a651 WatchSource:0}: Error finding container 55c7cb29ed4044abb3616655f924ef240e0a85e0510f904d15e60a9677d9a651: Status 404 returned error can't find the container with id 55c7cb29ed4044abb3616655f924ef240e0a85e0510f904d15e60a9677d9a651 Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.922573 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-8sb8s" event={"ID":"1e203487-144e-44d0-ba8e-f5374659484e","Type":"ContainerDied","Data":"74e5157a907a5cac1d21f7820bea92bc8afae3d1a9358592adf5964d8af5d585"} Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.922620 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-8sb8s" Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.927595 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c826858d-45cb-4ae3-8fa2-1531f9390480","Type":"ContainerStarted","Data":"8af3996ddb7090d07d17362f3923b3a06be028c2cd7798bd79695c31654fc4a9"} Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.930432 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"206b044a-6264-4c4e-89b9-7cacc4f30ac5","Type":"ContainerStarted","Data":"045672840c56ef9946cb205178286bcb1c4c5ad1cb0e96088a2973099a1e2caa"} Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.931728 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-brxh4" event={"ID":"b5f31813-ba17-4452-9b1a-ee7dcfac564e","Type":"ContainerStarted","Data":"6261c85436b24d45d624d8d0854e046566e5fc30778b276cf9f3d993ffef1cbb"} Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.934827 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9b27dc29-202f-4efb-9545-031eaa1d7da3","Type":"ContainerStarted","Data":"77b65e7dfc88feaf3be53c6a0db1737dacaaba242c76584ef96b55ec0c7882f2"} Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.936835 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a0dd3e13-e5ee-4315-ab7a-222c902192be","Type":"ContainerStarted","Data":"a9d4566a49a9d6f51e0d1c368fd734bd0cbb33d2225904a1d2e3bc6f34f28323"} Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.938113 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"644cd011-5c7b-4050-85e1-2194f5cd8d93","Type":"ContainerStarted","Data":"9c756d3206a98bcc761e1c0fa5a06c152c6c3fae18fcf3cc02c30fd7e7df8721"} Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.939548 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-8b4gw" event={"ID":"5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1","Type":"ContainerDied","Data":"01bea1c39138b6ab82bd6e8e487a6b43af40e8e81a065aba4ae70db1785ebc1c"} Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.939676 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-8b4gw" Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.942323 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"a27ef44f-2bff-4714-8018-a50f2f2de24a","Type":"ContainerStarted","Data":"55c7cb29ed4044abb3616655f924ef240e0a85e0510f904d15e60a9677d9a651"} Dec 03 06:44:39 crc kubenswrapper[4818]: I1203 06:44:39.943739 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kvg5t" event={"ID":"04bf523a-de6a-49a0-93ad-cca1e22939c6","Type":"ContainerStarted","Data":"55b1e5eab4d67166789ef3f3e15203da4b546022c2852f9906d24167245ec410"} Dec 03 06:44:40 crc kubenswrapper[4818]: I1203 06:44:40.047323 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8sb8s"] Dec 03 06:44:40 crc kubenswrapper[4818]: I1203 06:44:40.058365 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8sb8s"] Dec 03 06:44:40 crc kubenswrapper[4818]: I1203 06:44:40.081877 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8b4gw"] Dec 03 06:44:40 crc kubenswrapper[4818]: I1203 06:44:40.104962 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-8b4gw"] Dec 03 06:44:40 crc kubenswrapper[4818]: I1203 06:44:40.747211 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e203487-144e-44d0-ba8e-f5374659484e" path="/var/lib/kubelet/pods/1e203487-144e-44d0-ba8e-f5374659484e/volumes" Dec 03 06:44:40 crc kubenswrapper[4818]: I1203 06:44:40.747621 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1" path="/var/lib/kubelet/pods/5ef5eaf0-3a6f-45dc-a372-2f9f7fac17d1/volumes" Dec 03 06:44:42 crc kubenswrapper[4818]: I1203 06:44:42.969517 4818 generic.go:334] "Generic (PLEG): container finished" podID="206b044a-6264-4c4e-89b9-7cacc4f30ac5" containerID="045672840c56ef9946cb205178286bcb1c4c5ad1cb0e96088a2973099a1e2caa" exitCode=0 Dec 03 06:44:42 crc kubenswrapper[4818]: I1203 06:44:42.969600 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"206b044a-6264-4c4e-89b9-7cacc4f30ac5","Type":"ContainerDied","Data":"045672840c56ef9946cb205178286bcb1c4c5ad1cb0e96088a2973099a1e2caa"} Dec 03 06:44:42 crc kubenswrapper[4818]: I1203 06:44:42.972747 4818 generic.go:334] "Generic (PLEG): container finished" podID="a0dd3e13-e5ee-4315-ab7a-222c902192be" containerID="a9d4566a49a9d6f51e0d1c368fd734bd0cbb33d2225904a1d2e3bc6f34f28323" exitCode=0 Dec 03 06:44:42 crc kubenswrapper[4818]: I1203 06:44:42.972788 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a0dd3e13-e5ee-4315-ab7a-222c902192be","Type":"ContainerDied","Data":"a9d4566a49a9d6f51e0d1c368fd734bd0cbb33d2225904a1d2e3bc6f34f28323"} Dec 03 06:44:44 crc kubenswrapper[4818]: I1203 06:44:44.991943 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9b27dc29-202f-4efb-9545-031eaa1d7da3","Type":"ContainerStarted","Data":"85ff151be77120bde84814f5947b349ba940f8fcbf90b1dd72065086bc2358bf"} Dec 03 06:44:44 crc kubenswrapper[4818]: I1203 06:44:44.994717 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kvg5t" event={"ID":"04bf523a-de6a-49a0-93ad-cca1e22939c6","Type":"ContainerStarted","Data":"be8a2f2ad0a8c05daf72cbc1780889c77fdf4f213a40823ecb6d235ba21918d5"} Dec 03 06:44:44 crc kubenswrapper[4818]: I1203 06:44:44.994909 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-kvg5t" Dec 03 06:44:44 crc kubenswrapper[4818]: I1203 06:44:44.996424 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c826858d-45cb-4ae3-8fa2-1531f9390480","Type":"ContainerStarted","Data":"64afa146c566eb6ffe84dac922865604593394a24932450e362e42bf14a859e2"} Dec 03 06:44:44 crc kubenswrapper[4818]: I1203 06:44:44.996661 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 03 06:44:45 crc kubenswrapper[4818]: I1203 06:44:44.999757 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"206b044a-6264-4c4e-89b9-7cacc4f30ac5","Type":"ContainerStarted","Data":"907d99fe9fd66444365c9b6769ab30486fbc7a2e39b7b0cc3d2a2207aac9893d"} Dec 03 06:44:45 crc kubenswrapper[4818]: I1203 06:44:45.001909 4818 generic.go:334] "Generic (PLEG): container finished" podID="b5f31813-ba17-4452-9b1a-ee7dcfac564e" containerID="d69a8ed936be5b22100a4639d4fc8c79c320cf60d147b488fc9047757f160a03" exitCode=0 Dec 03 06:44:45 crc kubenswrapper[4818]: I1203 06:44:45.001983 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-brxh4" event={"ID":"b5f31813-ba17-4452-9b1a-ee7dcfac564e","Type":"ContainerDied","Data":"d69a8ed936be5b22100a4639d4fc8c79c320cf60d147b488fc9047757f160a03"} Dec 03 06:44:45 crc kubenswrapper[4818]: I1203 06:44:45.003890 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a0dd3e13-e5ee-4315-ab7a-222c902192be","Type":"ContainerStarted","Data":"bf69ca9e6c4e83bd84070ed1293a7a33613b5cf9bc58c2a70ec63f1afa2fe0bd"} Dec 03 06:44:45 crc kubenswrapper[4818]: I1203 06:44:45.007591 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"644cd011-5c7b-4050-85e1-2194f5cd8d93","Type":"ContainerStarted","Data":"f2fad1afc9124948c8ad10ceba58d3df7aaf314a58b99db8293a8339ae7a124a"} Dec 03 06:44:45 crc kubenswrapper[4818]: I1203 06:44:45.007720 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 06:44:45 crc kubenswrapper[4818]: I1203 06:44:45.011687 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"a27ef44f-2bff-4714-8018-a50f2f2de24a","Type":"ContainerStarted","Data":"c95ec4aa8bc402c338462df8c2ce32e35b995889e1cdb34eec21fe60f67f9204"} Dec 03 06:44:45 crc kubenswrapper[4818]: I1203 06:44:45.027489 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-kvg5t" podStartSLOduration=15.816501251 podStartE2EDuration="20.027474291s" podCreationTimestamp="2025-12-03 06:44:25 +0000 UTC" firstStartedPulling="2025-12-03 06:44:39.465467824 +0000 UTC m=+1037.157076576" lastFinishedPulling="2025-12-03 06:44:43.676440824 +0000 UTC m=+1041.368049616" observedRunningTime="2025-12-03 06:44:45.014597803 +0000 UTC m=+1042.706206555" watchObservedRunningTime="2025-12-03 06:44:45.027474291 +0000 UTC m=+1042.719083043" Dec 03 06:44:45 crc kubenswrapper[4818]: I1203 06:44:45.054131 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=14.037885368 podStartE2EDuration="29.05410935s" podCreationTimestamp="2025-12-03 06:44:16 +0000 UTC" firstStartedPulling="2025-12-03 06:44:24.01040878 +0000 UTC m=+1021.702017522" lastFinishedPulling="2025-12-03 06:44:39.026632762 +0000 UTC m=+1036.718241504" observedRunningTime="2025-12-03 06:44:45.053555746 +0000 UTC m=+1042.745164488" watchObservedRunningTime="2025-12-03 06:44:45.05410935 +0000 UTC m=+1042.745718102" Dec 03 06:44:45 crc kubenswrapper[4818]: I1203 06:44:45.076673 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=23.361157316 podStartE2EDuration="27.076637327s" podCreationTimestamp="2025-12-03 06:44:18 +0000 UTC" firstStartedPulling="2025-12-03 06:44:38.998190186 +0000 UTC m=+1036.689798938" lastFinishedPulling="2025-12-03 06:44:42.713670187 +0000 UTC m=+1040.405278949" observedRunningTime="2025-12-03 06:44:45.07232069 +0000 UTC m=+1042.763929442" watchObservedRunningTime="2025-12-03 06:44:45.076637327 +0000 UTC m=+1042.768246079" Dec 03 06:44:45 crc kubenswrapper[4818]: I1203 06:44:45.105058 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=19.273583319 podStartE2EDuration="24.105039589s" podCreationTimestamp="2025-12-03 06:44:21 +0000 UTC" firstStartedPulling="2025-12-03 06:44:38.954758569 +0000 UTC m=+1036.646367321" lastFinishedPulling="2025-12-03 06:44:43.786214839 +0000 UTC m=+1041.477823591" observedRunningTime="2025-12-03 06:44:45.092008637 +0000 UTC m=+1042.783617389" watchObservedRunningTime="2025-12-03 06:44:45.105039589 +0000 UTC m=+1042.796648341" Dec 03 06:44:45 crc kubenswrapper[4818]: I1203 06:44:45.117837 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=25.27567176 podStartE2EDuration="28.117806485s" podCreationTimestamp="2025-12-03 06:44:17 +0000 UTC" firstStartedPulling="2025-12-03 06:44:36.18461585 +0000 UTC m=+1033.876224602" lastFinishedPulling="2025-12-03 06:44:39.026750575 +0000 UTC m=+1036.718359327" observedRunningTime="2025-12-03 06:44:45.113900178 +0000 UTC m=+1042.805508930" watchObservedRunningTime="2025-12-03 06:44:45.117806485 +0000 UTC m=+1042.809415237" Dec 03 06:44:46 crc kubenswrapper[4818]: I1203 06:44:46.042760 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-brxh4" event={"ID":"b5f31813-ba17-4452-9b1a-ee7dcfac564e","Type":"ContainerStarted","Data":"37ee56d69bdf5bfcfa20f9e08e9f4dd01a76bf344f7d4f51c1bf63fa1163f312"} Dec 03 06:44:46 crc kubenswrapper[4818]: I1203 06:44:46.043131 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-brxh4" event={"ID":"b5f31813-ba17-4452-9b1a-ee7dcfac564e","Type":"ContainerStarted","Data":"a77135793a846888714c601d5ce698a852f9ea39e18a4b87cecf1e394dff7b2c"} Dec 03 06:44:46 crc kubenswrapper[4818]: I1203 06:44:46.043606 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:46 crc kubenswrapper[4818]: I1203 06:44:46.043662 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:44:46 crc kubenswrapper[4818]: I1203 06:44:46.075566 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-brxh4" podStartSLOduration=16.757057322 podStartE2EDuration="21.075544917s" podCreationTimestamp="2025-12-03 06:44:25 +0000 UTC" firstStartedPulling="2025-12-03 06:44:38.95073278 +0000 UTC m=+1036.642341532" lastFinishedPulling="2025-12-03 06:44:43.269220375 +0000 UTC m=+1040.960829127" observedRunningTime="2025-12-03 06:44:46.066108584 +0000 UTC m=+1043.757717356" watchObservedRunningTime="2025-12-03 06:44:46.075544917 +0000 UTC m=+1043.767153669" Dec 03 06:44:47 crc kubenswrapper[4818]: I1203 06:44:47.398003 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 03 06:44:47 crc kubenswrapper[4818]: I1203 06:44:47.398534 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 03 06:44:49 crc kubenswrapper[4818]: I1203 06:44:49.067753 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9b27dc29-202f-4efb-9545-031eaa1d7da3","Type":"ContainerStarted","Data":"85fdd833e992f38c9fd1c0cdbb80adad2e95ad8fb58726f1efea8e204cb9aeb4"} Dec 03 06:44:49 crc kubenswrapper[4818]: I1203 06:44:49.073504 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"a27ef44f-2bff-4714-8018-a50f2f2de24a","Type":"ContainerStarted","Data":"1914c9becc62e530f9ce1a1b320696e148ede8ad25559b02371c570bd19eb770"} Dec 03 06:44:49 crc kubenswrapper[4818]: I1203 06:44:49.097697 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=15.32479168 podStartE2EDuration="25.097677776s" podCreationTimestamp="2025-12-03 06:44:24 +0000 UTC" firstStartedPulling="2025-12-03 06:44:38.964105941 +0000 UTC m=+1036.655714733" lastFinishedPulling="2025-12-03 06:44:48.736992077 +0000 UTC m=+1046.428600829" observedRunningTime="2025-12-03 06:44:49.089650608 +0000 UTC m=+1046.781259370" watchObservedRunningTime="2025-12-03 06:44:49.097677776 +0000 UTC m=+1046.789286528" Dec 03 06:44:49 crc kubenswrapper[4818]: I1203 06:44:49.132199 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:49 crc kubenswrapper[4818]: I1203 06:44:49.132273 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 03 06:44:49 crc kubenswrapper[4818]: I1203 06:44:49.220881 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:49 crc kubenswrapper[4818]: I1203 06:44:49.246191 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 03 06:44:49 crc kubenswrapper[4818]: I1203 06:44:49.268884 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=13.42902293 podStartE2EDuration="22.268860699s" podCreationTimestamp="2025-12-03 06:44:27 +0000 UTC" firstStartedPulling="2025-12-03 06:44:39.895864456 +0000 UTC m=+1037.587473208" lastFinishedPulling="2025-12-03 06:44:48.735702225 +0000 UTC m=+1046.427310977" observedRunningTime="2025-12-03 06:44:49.119800453 +0000 UTC m=+1046.811409205" watchObservedRunningTime="2025-12-03 06:44:49.268860699 +0000 UTC m=+1046.960469451" Dec 03 06:44:50 crc kubenswrapper[4818]: I1203 06:44:50.221850 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:50 crc kubenswrapper[4818]: I1203 06:44:50.270910 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:50 crc kubenswrapper[4818]: I1203 06:44:50.388501 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:50 crc kubenswrapper[4818]: I1203 06:44:50.440459 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.091352 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"df9ab7c7-152b-4938-9ef5-d0588111b8a2","Type":"ContainerStarted","Data":"651b7927b0f2dd739e06ae0b1b47c6bf30fb2fa4bce7e634b77e016a5d7e4fce"} Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.092859 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0e751691-c660-4527-a9a6-61505f83c6ff","Type":"ContainerStarted","Data":"303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0"} Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.094228 4818 generic.go:334] "Generic (PLEG): container finished" podID="0b5ecd23-fa73-444e-a79c-166c4a25dff2" containerID="7d190cb5d11bb6790379c1b718a23058cf37d9a0cfc847c11f0221a59bed25d7" exitCode=0 Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.094268 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-d25tm" event={"ID":"0b5ecd23-fa73-444e-a79c-166c4a25dff2","Type":"ContainerDied","Data":"7d190cb5d11bb6790379c1b718a23058cf37d9a0cfc847c11f0221a59bed25d7"} Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.094853 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.161379 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.175902 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.397086 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.475294 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f2vzn"] Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.525260 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-6sprz"] Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.526554 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.558782 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be50273d-b600-4461-8c58-69cf20892e97-config\") pod \"dnsmasq-dns-7cb5889db5-6sprz\" (UID: \"be50273d-b600-4461-8c58-69cf20892e97\") " pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.558837 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be50273d-b600-4461-8c58-69cf20892e97-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-6sprz\" (UID: \"be50273d-b600-4461-8c58-69cf20892e97\") " pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.558878 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxm44\" (UniqueName: \"kubernetes.io/projected/be50273d-b600-4461-8c58-69cf20892e97-kube-api-access-pxm44\") pod \"dnsmasq-dns-7cb5889db5-6sprz\" (UID: \"be50273d-b600-4461-8c58-69cf20892e97\") " pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.559161 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-6sprz"] Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.653257 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-d25tm"] Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.660845 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxm44\" (UniqueName: \"kubernetes.io/projected/be50273d-b600-4461-8c58-69cf20892e97-kube-api-access-pxm44\") pod \"dnsmasq-dns-7cb5889db5-6sprz\" (UID: \"be50273d-b600-4461-8c58-69cf20892e97\") " pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.660990 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be50273d-b600-4461-8c58-69cf20892e97-config\") pod \"dnsmasq-dns-7cb5889db5-6sprz\" (UID: \"be50273d-b600-4461-8c58-69cf20892e97\") " pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.661014 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be50273d-b600-4461-8c58-69cf20892e97-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-6sprz\" (UID: \"be50273d-b600-4461-8c58-69cf20892e97\") " pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.662102 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be50273d-b600-4461-8c58-69cf20892e97-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-6sprz\" (UID: \"be50273d-b600-4461-8c58-69cf20892e97\") " pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.662872 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be50273d-b600-4461-8c58-69cf20892e97-config\") pod \"dnsmasq-dns-7cb5889db5-6sprz\" (UID: \"be50273d-b600-4461-8c58-69cf20892e97\") " pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.711510 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-srrh4"] Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.723298 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.728192 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxm44\" (UniqueName: \"kubernetes.io/projected/be50273d-b600-4461-8c58-69cf20892e97-kube-api-access-pxm44\") pod \"dnsmasq-dns-7cb5889db5-6sprz\" (UID: \"be50273d-b600-4461-8c58-69cf20892e97\") " pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.733677 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.741638 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-srrh4"] Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.756051 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-sww4m"] Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.757688 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-sww4m" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.763419 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-sww4m\" (UID: \"091316da-41cb-491b-80b6-dfe633c68c06\") " pod="openstack/dnsmasq-dns-6c89d5d749-sww4m" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.763466 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/997e3f50-bc25-4436-b03d-5e98770df25f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.763491 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nfm9\" (UniqueName: \"kubernetes.io/projected/091316da-41cb-491b-80b6-dfe633c68c06-kube-api-access-6nfm9\") pod \"dnsmasq-dns-6c89d5d749-sww4m\" (UID: \"091316da-41cb-491b-80b6-dfe633c68c06\") " pod="openstack/dnsmasq-dns-6c89d5d749-sww4m" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.763509 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-sww4m\" (UID: \"091316da-41cb-491b-80b6-dfe633c68c06\") " pod="openstack/dnsmasq-dns-6c89d5d749-sww4m" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.763733 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/997e3f50-bc25-4436-b03d-5e98770df25f-config\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.763793 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs2kk\" (UniqueName: \"kubernetes.io/projected/997e3f50-bc25-4436-b03d-5e98770df25f-kube-api-access-fs2kk\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.763911 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-config\") pod \"dnsmasq-dns-6c89d5d749-sww4m\" (UID: \"091316da-41cb-491b-80b6-dfe633c68c06\") " pod="openstack/dnsmasq-dns-6c89d5d749-sww4m" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.764209 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/997e3f50-bc25-4436-b03d-5e98770df25f-combined-ca-bundle\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.764427 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/997e3f50-bc25-4436-b03d-5e98770df25f-ovs-rundir\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.764468 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/997e3f50-bc25-4436-b03d-5e98770df25f-ovn-rundir\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.777925 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.801526 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-sww4m"] Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.815335 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.848001 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.849311 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.856208 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.856317 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.856217 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.857917 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-vhtbh" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.876712 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.879832 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/997e3f50-bc25-4436-b03d-5e98770df25f-combined-ca-bundle\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.879949 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/997e3f50-bc25-4436-b03d-5e98770df25f-ovs-rundir\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.880025 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/997e3f50-bc25-4436-b03d-5e98770df25f-ovn-rundir\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.880101 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-sww4m\" (UID: \"091316da-41cb-491b-80b6-dfe633c68c06\") " pod="openstack/dnsmasq-dns-6c89d5d749-sww4m" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.880173 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/997e3f50-bc25-4436-b03d-5e98770df25f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.880243 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nfm9\" (UniqueName: \"kubernetes.io/projected/091316da-41cb-491b-80b6-dfe633c68c06-kube-api-access-6nfm9\") pod \"dnsmasq-dns-6c89d5d749-sww4m\" (UID: \"091316da-41cb-491b-80b6-dfe633c68c06\") " pod="openstack/dnsmasq-dns-6c89d5d749-sww4m" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.880309 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-sww4m\" (UID: \"091316da-41cb-491b-80b6-dfe633c68c06\") " pod="openstack/dnsmasq-dns-6c89d5d749-sww4m" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.880399 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/997e3f50-bc25-4436-b03d-5e98770df25f-config\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.880470 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs2kk\" (UniqueName: \"kubernetes.io/projected/997e3f50-bc25-4436-b03d-5e98770df25f-kube-api-access-fs2kk\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.880545 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-config\") pod \"dnsmasq-dns-6c89d5d749-sww4m\" (UID: \"091316da-41cb-491b-80b6-dfe633c68c06\") " pod="openstack/dnsmasq-dns-6c89d5d749-sww4m" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.881700 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-config\") pod \"dnsmasq-dns-6c89d5d749-sww4m\" (UID: \"091316da-41cb-491b-80b6-dfe633c68c06\") " pod="openstack/dnsmasq-dns-6c89d5d749-sww4m" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.881967 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/997e3f50-bc25-4436-b03d-5e98770df25f-ovs-rundir\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.882080 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/997e3f50-bc25-4436-b03d-5e98770df25f-ovn-rundir\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.882636 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-sww4m\" (UID: \"091316da-41cb-491b-80b6-dfe633c68c06\") " pod="openstack/dnsmasq-dns-6c89d5d749-sww4m" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.883915 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-sww4m\" (UID: \"091316da-41cb-491b-80b6-dfe633c68c06\") " pod="openstack/dnsmasq-dns-6c89d5d749-sww4m" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.883984 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/997e3f50-bc25-4436-b03d-5e98770df25f-config\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.898693 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/997e3f50-bc25-4436-b03d-5e98770df25f-combined-ca-bundle\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.905055 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/997e3f50-bc25-4436-b03d-5e98770df25f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.923893 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.937615 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-sww4m"] Dec 03 06:44:51 crc kubenswrapper[4818]: E1203 06:44:51.938638 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-6nfm9], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-6c89d5d749-sww4m" podUID="091316da-41cb-491b-80b6-dfe633c68c06" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.944705 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nfm9\" (UniqueName: \"kubernetes.io/projected/091316da-41cb-491b-80b6-dfe633c68c06-kube-api-access-6nfm9\") pod \"dnsmasq-dns-6c89d5d749-sww4m\" (UID: \"091316da-41cb-491b-80b6-dfe633c68c06\") " pod="openstack/dnsmasq-dns-6c89d5d749-sww4m" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.960149 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-8p9jz"] Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.961384 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.981685 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7l68\" (UniqueName: \"kubernetes.io/projected/8df18ace-e467-4e73-9304-3ea16df510e8-kube-api-access-c7l68\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.981780 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8df18ace-e467-4e73-9304-3ea16df510e8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.981804 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df18ace-e467-4e73-9304-3ea16df510e8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.981852 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8df18ace-e467-4e73-9304-3ea16df510e8-config\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.981925 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8df18ace-e467-4e73-9304-3ea16df510e8-scripts\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.981939 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8df18ace-e467-4e73-9304-3ea16df510e8-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.981971 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8df18ace-e467-4e73-9304-3ea16df510e8-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.984952 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.990640 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-8p9jz"] Dec 03 06:44:51 crc kubenswrapper[4818]: I1203 06:44:51.994080 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs2kk\" (UniqueName: \"kubernetes.io/projected/997e3f50-bc25-4436-b03d-5e98770df25f-kube-api-access-fs2kk\") pod \"ovn-controller-metrics-srrh4\" (UID: \"997e3f50-bc25-4436-b03d-5e98770df25f\") " pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.081805 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-srrh4" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.083907 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8df18ace-e467-4e73-9304-3ea16df510e8-scripts\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.083955 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8df18ace-e467-4e73-9304-3ea16df510e8-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.084002 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8df18ace-e467-4e73-9304-3ea16df510e8-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.084030 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-8p9jz\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.084091 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82hjq\" (UniqueName: \"kubernetes.io/projected/6bf0813f-9f49-43b7-9e9a-8712e6a86159-kube-api-access-82hjq\") pod \"dnsmasq-dns-698758b865-8p9jz\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.084112 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7l68\" (UniqueName: \"kubernetes.io/projected/8df18ace-e467-4e73-9304-3ea16df510e8-kube-api-access-c7l68\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.084142 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-config\") pod \"dnsmasq-dns-698758b865-8p9jz\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.084169 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-8p9jz\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.084187 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8df18ace-e467-4e73-9304-3ea16df510e8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.084214 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df18ace-e467-4e73-9304-3ea16df510e8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.084236 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-dns-svc\") pod \"dnsmasq-dns-698758b865-8p9jz\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.084262 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8df18ace-e467-4e73-9304-3ea16df510e8-config\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.084838 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8df18ace-e467-4e73-9304-3ea16df510e8-scripts\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.089494 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8df18ace-e467-4e73-9304-3ea16df510e8-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.089840 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8df18ace-e467-4e73-9304-3ea16df510e8-config\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.090038 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-f2vzn" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.098014 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df18ace-e467-4e73-9304-3ea16df510e8-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.098239 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8df18ace-e467-4e73-9304-3ea16df510e8-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.098340 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8df18ace-e467-4e73-9304-3ea16df510e8-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.115063 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7l68\" (UniqueName: \"kubernetes.io/projected/8df18ace-e467-4e73-9304-3ea16df510e8-kube-api-access-c7l68\") pod \"ovn-northd-0\" (UID: \"8df18ace-e467-4e73-9304-3ea16df510e8\") " pod="openstack/ovn-northd-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.146521 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.162624 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-d25tm" podUID="0b5ecd23-fa73-444e-a79c-166c4a25dff2" containerName="dnsmasq-dns" containerID="cri-o://871491b91338481cc0c5af41a8ce042e475ec1cc84c51ebacd4f01c47b63edc5" gracePeriod=10 Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.162699 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-d25tm" event={"ID":"0b5ecd23-fa73-444e-a79c-166c4a25dff2","Type":"ContainerStarted","Data":"871491b91338481cc0c5af41a8ce042e475ec1cc84c51ebacd4f01c47b63edc5"} Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.162783 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-d25tm" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.187370 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11a1627a-c749-4212-a52b-61e284a03211-dns-svc\") pod \"11a1627a-c749-4212-a52b-61e284a03211\" (UID: \"11a1627a-c749-4212-a52b-61e284a03211\") " Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.187599 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xs9fk\" (UniqueName: \"kubernetes.io/projected/11a1627a-c749-4212-a52b-61e284a03211-kube-api-access-xs9fk\") pod \"11a1627a-c749-4212-a52b-61e284a03211\" (UID: \"11a1627a-c749-4212-a52b-61e284a03211\") " Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.187719 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11a1627a-c749-4212-a52b-61e284a03211-config\") pod \"11a1627a-c749-4212-a52b-61e284a03211\" (UID: \"11a1627a-c749-4212-a52b-61e284a03211\") " Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.190167 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-8p9jz\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.190265 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82hjq\" (UniqueName: \"kubernetes.io/projected/6bf0813f-9f49-43b7-9e9a-8712e6a86159-kube-api-access-82hjq\") pod \"dnsmasq-dns-698758b865-8p9jz\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.190310 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-config\") pod \"dnsmasq-dns-698758b865-8p9jz\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.190345 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-8p9jz\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.190404 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-dns-svc\") pod \"dnsmasq-dns-698758b865-8p9jz\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.193655 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-8p9jz\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.194571 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-8p9jz\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.195406 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-config\") pod \"dnsmasq-dns-698758b865-8p9jz\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.196134 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11a1627a-c749-4212-a52b-61e284a03211-config" (OuterVolumeSpecName: "config") pod "11a1627a-c749-4212-a52b-61e284a03211" (UID: "11a1627a-c749-4212-a52b-61e284a03211"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.197068 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f2vzn" event={"ID":"11a1627a-c749-4212-a52b-61e284a03211","Type":"ContainerDied","Data":"d1ca1bcdc4619647d1bb88cb5dfa3fa5e04781cb08993f2e994add6d3a35af35"} Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.197080 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-dns-svc\") pod \"dnsmasq-dns-698758b865-8p9jz\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.197162 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-f2vzn" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.197532 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-sww4m" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.198548 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11a1627a-c749-4212-a52b-61e284a03211-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "11a1627a-c749-4212-a52b-61e284a03211" (UID: "11a1627a-c749-4212-a52b-61e284a03211"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.212370 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11a1627a-c749-4212-a52b-61e284a03211-kube-api-access-xs9fk" (OuterVolumeSpecName: "kube-api-access-xs9fk") pod "11a1627a-c749-4212-a52b-61e284a03211" (UID: "11a1627a-c749-4212-a52b-61e284a03211"). InnerVolumeSpecName "kube-api-access-xs9fk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.222226 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82hjq\" (UniqueName: \"kubernetes.io/projected/6bf0813f-9f49-43b7-9e9a-8712e6a86159-kube-api-access-82hjq\") pod \"dnsmasq-dns-698758b865-8p9jz\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.225033 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-sww4m" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.245014 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-d25tm" podStartSLOduration=3.005065064 podStartE2EDuration="38.244985011s" podCreationTimestamp="2025-12-03 06:44:14 +0000 UTC" firstStartedPulling="2025-12-03 06:44:15.150410013 +0000 UTC m=+1012.842018765" lastFinishedPulling="2025-12-03 06:44:50.39032996 +0000 UTC m=+1048.081938712" observedRunningTime="2025-12-03 06:44:52.211294048 +0000 UTC m=+1049.902902800" watchObservedRunningTime="2025-12-03 06:44:52.244985011 +0000 UTC m=+1049.936593763" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.292790 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-ovsdbserver-sb\") pod \"091316da-41cb-491b-80b6-dfe633c68c06\" (UID: \"091316da-41cb-491b-80b6-dfe633c68c06\") " Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.292916 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-dns-svc\") pod \"091316da-41cb-491b-80b6-dfe633c68c06\" (UID: \"091316da-41cb-491b-80b6-dfe633c68c06\") " Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.293041 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nfm9\" (UniqueName: \"kubernetes.io/projected/091316da-41cb-491b-80b6-dfe633c68c06-kube-api-access-6nfm9\") pod \"091316da-41cb-491b-80b6-dfe633c68c06\" (UID: \"091316da-41cb-491b-80b6-dfe633c68c06\") " Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.293223 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-config\") pod \"091316da-41cb-491b-80b6-dfe633c68c06\" (UID: \"091316da-41cb-491b-80b6-dfe633c68c06\") " Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.293458 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "091316da-41cb-491b-80b6-dfe633c68c06" (UID: "091316da-41cb-491b-80b6-dfe633c68c06"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.293907 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11a1627a-c749-4212-a52b-61e284a03211-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.293923 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11a1627a-c749-4212-a52b-61e284a03211-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.293933 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.293947 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xs9fk\" (UniqueName: \"kubernetes.io/projected/11a1627a-c749-4212-a52b-61e284a03211-kube-api-access-xs9fk\") on node \"crc\" DevicePath \"\"" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.294606 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "091316da-41cb-491b-80b6-dfe633c68c06" (UID: "091316da-41cb-491b-80b6-dfe633c68c06"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.296840 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-config" (OuterVolumeSpecName: "config") pod "091316da-41cb-491b-80b6-dfe633c68c06" (UID: "091316da-41cb-491b-80b6-dfe633c68c06"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.337794 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.382933 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.394983 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.395012 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/091316da-41cb-491b-80b6-dfe633c68c06-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.415270 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/091316da-41cb-491b-80b6-dfe633c68c06-kube-api-access-6nfm9" (OuterVolumeSpecName: "kube-api-access-6nfm9") pod "091316da-41cb-491b-80b6-dfe633c68c06" (UID: "091316da-41cb-491b-80b6-dfe633c68c06"). InnerVolumeSpecName "kube-api-access-6nfm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.495952 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nfm9\" (UniqueName: \"kubernetes.io/projected/091316da-41cb-491b-80b6-dfe633c68c06-kube-api-access-6nfm9\") on node \"crc\" DevicePath \"\"" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.511737 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-6sprz"] Dec 03 06:44:52 crc kubenswrapper[4818]: W1203 06:44:52.536931 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe50273d_b600_4461_8c58_69cf20892e97.slice/crio-79e5b4a6e916af18aebb3e3cb1450c83d9b354e0df16104b21123d9c75cdc579 WatchSource:0}: Error finding container 79e5b4a6e916af18aebb3e3cb1450c83d9b354e0df16104b21123d9c75cdc579: Status 404 returned error can't find the container with id 79e5b4a6e916af18aebb3e3cb1450c83d9b354e0df16104b21123d9c75cdc579 Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.647283 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-srrh4"] Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.654791 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f2vzn"] Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.659155 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f2vzn"] Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.663183 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.668490 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.673993 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.674293 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.674481 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.674605 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-tzk5t" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.706525 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.804951 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.804996 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-cache\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.805052 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rmw4\" (UniqueName: \"kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-kube-api-access-6rmw4\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.805107 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-lock\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.805136 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.825859 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11a1627a-c749-4212-a52b-61e284a03211" path="/var/lib/kubelet/pods/11a1627a-c749-4212-a52b-61e284a03211/volumes" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.840667 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-8p9jz"] Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.906900 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.906941 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-cache\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.906993 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rmw4\" (UniqueName: \"kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-kube-api-access-6rmw4\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.907047 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-lock\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.907071 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:44:52 crc kubenswrapper[4818]: E1203 06:44:52.907330 4818 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 06:44:52 crc kubenswrapper[4818]: E1203 06:44:52.907354 4818 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 06:44:52 crc kubenswrapper[4818]: E1203 06:44:52.907409 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift podName:520a6b29-d2d4-46e4-a5ed-989f2fa1df1f nodeName:}" failed. No retries permitted until 2025-12-03 06:44:53.40738665 +0000 UTC m=+1051.098995392 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift") pod "swift-storage-0" (UID: "520a6b29-d2d4-46e4-a5ed-989f2fa1df1f") : configmap "swift-ring-files" not found Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.908084 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-lock\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.908366 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.908375 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-cache\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.943774 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rmw4\" (UniqueName: \"kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-kube-api-access-6rmw4\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:44:52 crc kubenswrapper[4818]: I1203 06:44:52.947050 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:44:53 crc kubenswrapper[4818]: I1203 06:44:53.083033 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 06:44:53 crc kubenswrapper[4818]: I1203 06:44:53.211629 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-srrh4" event={"ID":"997e3f50-bc25-4436-b03d-5e98770df25f","Type":"ContainerStarted","Data":"2e726f36fed3a252fae1269f680cbfea75e2f719b8ea57f9e50f833fda392ef1"} Dec 03 06:44:53 crc kubenswrapper[4818]: I1203 06:44:53.218860 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" event={"ID":"be50273d-b600-4461-8c58-69cf20892e97","Type":"ContainerStarted","Data":"79e5b4a6e916af18aebb3e3cb1450c83d9b354e0df16104b21123d9c75cdc579"} Dec 03 06:44:53 crc kubenswrapper[4818]: I1203 06:44:53.223077 4818 generic.go:334] "Generic (PLEG): container finished" podID="0b5ecd23-fa73-444e-a79c-166c4a25dff2" containerID="871491b91338481cc0c5af41a8ce042e475ec1cc84c51ebacd4f01c47b63edc5" exitCode=0 Dec 03 06:44:53 crc kubenswrapper[4818]: I1203 06:44:53.223176 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-d25tm" event={"ID":"0b5ecd23-fa73-444e-a79c-166c4a25dff2","Type":"ContainerDied","Data":"871491b91338481cc0c5af41a8ce042e475ec1cc84c51ebacd4f01c47b63edc5"} Dec 03 06:44:53 crc kubenswrapper[4818]: I1203 06:44:53.224451 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8df18ace-e467-4e73-9304-3ea16df510e8","Type":"ContainerStarted","Data":"c160d5489a54eb96008b50d48eb24f9054b21575d36f963c5bd187eadf5cf375"} Dec 03 06:44:53 crc kubenswrapper[4818]: I1203 06:44:53.228249 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-8p9jz" event={"ID":"6bf0813f-9f49-43b7-9e9a-8712e6a86159","Type":"ContainerStarted","Data":"981950291dede3a915ad46ea7bce9ce057bb3e581fd0412c1149d4e81019c50d"} Dec 03 06:44:53 crc kubenswrapper[4818]: I1203 06:44:53.228397 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-sww4m" Dec 03 06:44:53 crc kubenswrapper[4818]: I1203 06:44:53.278051 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-sww4m"] Dec 03 06:44:53 crc kubenswrapper[4818]: I1203 06:44:53.289943 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-sww4m"] Dec 03 06:44:53 crc kubenswrapper[4818]: I1203 06:44:53.423704 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:44:53 crc kubenswrapper[4818]: E1203 06:44:53.424007 4818 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 06:44:53 crc kubenswrapper[4818]: E1203 06:44:53.424354 4818 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 06:44:53 crc kubenswrapper[4818]: E1203 06:44:53.424413 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift podName:520a6b29-d2d4-46e4-a5ed-989f2fa1df1f nodeName:}" failed. No retries permitted until 2025-12-03 06:44:54.424396385 +0000 UTC m=+1052.116005137 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift") pod "swift-storage-0" (UID: "520a6b29-d2d4-46e4-a5ed-989f2fa1df1f") : configmap "swift-ring-files" not found Dec 03 06:44:54 crc kubenswrapper[4818]: I1203 06:44:54.441769 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:44:54 crc kubenswrapper[4818]: E1203 06:44:54.441974 4818 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 06:44:54 crc kubenswrapper[4818]: E1203 06:44:54.442001 4818 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 06:44:54 crc kubenswrapper[4818]: E1203 06:44:54.442073 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift podName:520a6b29-d2d4-46e4-a5ed-989f2fa1df1f nodeName:}" failed. No retries permitted until 2025-12-03 06:44:56.442052869 +0000 UTC m=+1054.133661621 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift") pod "swift-storage-0" (UID: "520a6b29-d2d4-46e4-a5ed-989f2fa1df1f") : configmap "swift-ring-files" not found Dec 03 06:44:54 crc kubenswrapper[4818]: I1203 06:44:54.749189 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="091316da-41cb-491b-80b6-dfe633c68c06" path="/var/lib/kubelet/pods/091316da-41cb-491b-80b6-dfe633c68c06/volumes" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.486739 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:44:56 crc kubenswrapper[4818]: E1203 06:44:56.486982 4818 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 06:44:56 crc kubenswrapper[4818]: E1203 06:44:56.487350 4818 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 06:44:56 crc kubenswrapper[4818]: E1203 06:44:56.487421 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift podName:520a6b29-d2d4-46e4-a5ed-989f2fa1df1f nodeName:}" failed. No retries permitted until 2025-12-03 06:45:00.487397414 +0000 UTC m=+1058.179006196 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift") pod "swift-storage-0" (UID: "520a6b29-d2d4-46e4-a5ed-989f2fa1df1f") : configmap "swift-ring-files" not found Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.704520 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-pkxxx"] Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.706062 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.708371 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.708666 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.709268 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.724499 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-pkxxx"] Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.794005 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz5cg\" (UniqueName: \"kubernetes.io/projected/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-kube-api-access-wz5cg\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.794207 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-etc-swift\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.794274 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-combined-ca-bundle\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.794317 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-scripts\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.794364 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-ring-data-devices\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.794388 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-swiftconf\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.794474 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-dispersionconf\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.895383 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-dispersionconf\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.895434 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz5cg\" (UniqueName: \"kubernetes.io/projected/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-kube-api-access-wz5cg\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.895481 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-etc-swift\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.895516 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-combined-ca-bundle\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.895536 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-scripts\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.895572 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-ring-data-devices\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.895586 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-swiftconf\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.897178 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-etc-swift\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.897709 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-ring-data-devices\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.897775 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-scripts\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.903730 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-swiftconf\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.906536 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-combined-ca-bundle\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.906876 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-dispersionconf\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:56 crc kubenswrapper[4818]: I1203 06:44:56.914569 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz5cg\" (UniqueName: \"kubernetes.io/projected/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-kube-api-access-wz5cg\") pod \"swift-ring-rebalance-pkxxx\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:57 crc kubenswrapper[4818]: I1203 06:44:57.043941 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:44:57 crc kubenswrapper[4818]: I1203 06:44:57.519009 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-pkxxx"] Dec 03 06:44:57 crc kubenswrapper[4818]: W1203 06:44:57.530864 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod337c7d6b_71fc_4044_98ec_de8bfbf5dd6f.slice/crio-576d6118befbfed8152e7db32e7ca031f9797b3b2c6dc18ebaeb97bcba5b89df WatchSource:0}: Error finding container 576d6118befbfed8152e7db32e7ca031f9797b3b2c6dc18ebaeb97bcba5b89df: Status 404 returned error can't find the container with id 576d6118befbfed8152e7db32e7ca031f9797b3b2c6dc18ebaeb97bcba5b89df Dec 03 06:44:58 crc kubenswrapper[4818]: I1203 06:44:58.271298 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-pkxxx" event={"ID":"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f","Type":"ContainerStarted","Data":"576d6118befbfed8152e7db32e7ca031f9797b3b2c6dc18ebaeb97bcba5b89df"} Dec 03 06:44:58 crc kubenswrapper[4818]: I1203 06:44:58.900936 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-15a6-account-create-update-rvmpz"] Dec 03 06:44:58 crc kubenswrapper[4818]: I1203 06:44:58.904166 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-15a6-account-create-update-rvmpz" Dec 03 06:44:58 crc kubenswrapper[4818]: I1203 06:44:58.908261 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 03 06:44:58 crc kubenswrapper[4818]: I1203 06:44:58.922706 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-15a6-account-create-update-rvmpz"] Dec 03 06:44:58 crc kubenswrapper[4818]: I1203 06:44:58.932032 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-d25tm" Dec 03 06:44:58 crc kubenswrapper[4818]: I1203 06:44:58.938863 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-x7rhg"] Dec 03 06:44:58 crc kubenswrapper[4818]: E1203 06:44:58.939453 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b5ecd23-fa73-444e-a79c-166c4a25dff2" containerName="init" Dec 03 06:44:58 crc kubenswrapper[4818]: I1203 06:44:58.939472 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b5ecd23-fa73-444e-a79c-166c4a25dff2" containerName="init" Dec 03 06:44:58 crc kubenswrapper[4818]: E1203 06:44:58.939551 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b5ecd23-fa73-444e-a79c-166c4a25dff2" containerName="dnsmasq-dns" Dec 03 06:44:58 crc kubenswrapper[4818]: I1203 06:44:58.939573 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b5ecd23-fa73-444e-a79c-166c4a25dff2" containerName="dnsmasq-dns" Dec 03 06:44:58 crc kubenswrapper[4818]: I1203 06:44:58.939849 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b5ecd23-fa73-444e-a79c-166c4a25dff2" containerName="dnsmasq-dns" Dec 03 06:44:58 crc kubenswrapper[4818]: I1203 06:44:58.940848 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-x7rhg" Dec 03 06:44:58 crc kubenswrapper[4818]: I1203 06:44:58.942133 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/619d6969-7ea3-4680-ba99-061f79fcadcc-operator-scripts\") pod \"keystone-15a6-account-create-update-rvmpz\" (UID: \"619d6969-7ea3-4680-ba99-061f79fcadcc\") " pod="openstack/keystone-15a6-account-create-update-rvmpz" Dec 03 06:44:58 crc kubenswrapper[4818]: I1203 06:44:58.942294 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sxtb\" (UniqueName: \"kubernetes.io/projected/619d6969-7ea3-4680-ba99-061f79fcadcc-kube-api-access-5sxtb\") pod \"keystone-15a6-account-create-update-rvmpz\" (UID: \"619d6969-7ea3-4680-ba99-061f79fcadcc\") " pod="openstack/keystone-15a6-account-create-update-rvmpz" Dec 03 06:44:58 crc kubenswrapper[4818]: I1203 06:44:58.945203 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-x7rhg"] Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.043242 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b5ecd23-fa73-444e-a79c-166c4a25dff2-dns-svc\") pod \"0b5ecd23-fa73-444e-a79c-166c4a25dff2\" (UID: \"0b5ecd23-fa73-444e-a79c-166c4a25dff2\") " Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.043339 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b5ecd23-fa73-444e-a79c-166c4a25dff2-config\") pod \"0b5ecd23-fa73-444e-a79c-166c4a25dff2\" (UID: \"0b5ecd23-fa73-444e-a79c-166c4a25dff2\") " Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.043569 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zk8ks\" (UniqueName: \"kubernetes.io/projected/0b5ecd23-fa73-444e-a79c-166c4a25dff2-kube-api-access-zk8ks\") pod \"0b5ecd23-fa73-444e-a79c-166c4a25dff2\" (UID: \"0b5ecd23-fa73-444e-a79c-166c4a25dff2\") " Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.044216 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/619d6969-7ea3-4680-ba99-061f79fcadcc-operator-scripts\") pod \"keystone-15a6-account-create-update-rvmpz\" (UID: \"619d6969-7ea3-4680-ba99-061f79fcadcc\") " pod="openstack/keystone-15a6-account-create-update-rvmpz" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.044277 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dxww\" (UniqueName: \"kubernetes.io/projected/19cc3001-7257-4267-a3e2-822ebf77b969-kube-api-access-8dxww\") pod \"keystone-db-create-x7rhg\" (UID: \"19cc3001-7257-4267-a3e2-822ebf77b969\") " pod="openstack/keystone-db-create-x7rhg" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.044333 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19cc3001-7257-4267-a3e2-822ebf77b969-operator-scripts\") pod \"keystone-db-create-x7rhg\" (UID: \"19cc3001-7257-4267-a3e2-822ebf77b969\") " pod="openstack/keystone-db-create-x7rhg" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.044380 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sxtb\" (UniqueName: \"kubernetes.io/projected/619d6969-7ea3-4680-ba99-061f79fcadcc-kube-api-access-5sxtb\") pod \"keystone-15a6-account-create-update-rvmpz\" (UID: \"619d6969-7ea3-4680-ba99-061f79fcadcc\") " pod="openstack/keystone-15a6-account-create-update-rvmpz" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.045603 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/619d6969-7ea3-4680-ba99-061f79fcadcc-operator-scripts\") pod \"keystone-15a6-account-create-update-rvmpz\" (UID: \"619d6969-7ea3-4680-ba99-061f79fcadcc\") " pod="openstack/keystone-15a6-account-create-update-rvmpz" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.048769 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b5ecd23-fa73-444e-a79c-166c4a25dff2-kube-api-access-zk8ks" (OuterVolumeSpecName: "kube-api-access-zk8ks") pod "0b5ecd23-fa73-444e-a79c-166c4a25dff2" (UID: "0b5ecd23-fa73-444e-a79c-166c4a25dff2"). InnerVolumeSpecName "kube-api-access-zk8ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.064133 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sxtb\" (UniqueName: \"kubernetes.io/projected/619d6969-7ea3-4680-ba99-061f79fcadcc-kube-api-access-5sxtb\") pod \"keystone-15a6-account-create-update-rvmpz\" (UID: \"619d6969-7ea3-4680-ba99-061f79fcadcc\") " pod="openstack/keystone-15a6-account-create-update-rvmpz" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.082223 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b5ecd23-fa73-444e-a79c-166c4a25dff2-config" (OuterVolumeSpecName: "config") pod "0b5ecd23-fa73-444e-a79c-166c4a25dff2" (UID: "0b5ecd23-fa73-444e-a79c-166c4a25dff2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.084936 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b5ecd23-fa73-444e-a79c-166c4a25dff2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0b5ecd23-fa73-444e-a79c-166c4a25dff2" (UID: "0b5ecd23-fa73-444e-a79c-166c4a25dff2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.127537 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-md6rj"] Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.129613 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-md6rj" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.139045 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-md6rj"] Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.151057 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dxww\" (UniqueName: \"kubernetes.io/projected/19cc3001-7257-4267-a3e2-822ebf77b969-kube-api-access-8dxww\") pod \"keystone-db-create-x7rhg\" (UID: \"19cc3001-7257-4267-a3e2-822ebf77b969\") " pod="openstack/keystone-db-create-x7rhg" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.151118 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19cc3001-7257-4267-a3e2-822ebf77b969-operator-scripts\") pod \"keystone-db-create-x7rhg\" (UID: \"19cc3001-7257-4267-a3e2-822ebf77b969\") " pod="openstack/keystone-db-create-x7rhg" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.151244 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b5ecd23-fa73-444e-a79c-166c4a25dff2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.151254 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b5ecd23-fa73-444e-a79c-166c4a25dff2-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.151264 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zk8ks\" (UniqueName: \"kubernetes.io/projected/0b5ecd23-fa73-444e-a79c-166c4a25dff2-kube-api-access-zk8ks\") on node \"crc\" DevicePath \"\"" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.153455 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19cc3001-7257-4267-a3e2-822ebf77b969-operator-scripts\") pod \"keystone-db-create-x7rhg\" (UID: \"19cc3001-7257-4267-a3e2-822ebf77b969\") " pod="openstack/keystone-db-create-x7rhg" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.167354 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dxww\" (UniqueName: \"kubernetes.io/projected/19cc3001-7257-4267-a3e2-822ebf77b969-kube-api-access-8dxww\") pod \"keystone-db-create-x7rhg\" (UID: \"19cc3001-7257-4267-a3e2-822ebf77b969\") " pod="openstack/keystone-db-create-x7rhg" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.242842 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9db7-account-create-update-6n5md"] Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.244477 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9db7-account-create-update-6n5md" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.249175 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.253021 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg777\" (UniqueName: \"kubernetes.io/projected/9bb6d069-8823-4c5c-be08-d9bbb831e977-kube-api-access-vg777\") pod \"placement-db-create-md6rj\" (UID: \"9bb6d069-8823-4c5c-be08-d9bbb831e977\") " pod="openstack/placement-db-create-md6rj" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.253155 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9bb6d069-8823-4c5c-be08-d9bbb831e977-operator-scripts\") pod \"placement-db-create-md6rj\" (UID: \"9bb6d069-8823-4c5c-be08-d9bbb831e977\") " pod="openstack/placement-db-create-md6rj" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.253254 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-15a6-account-create-update-rvmpz" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.264469 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-x7rhg" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.276615 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9db7-account-create-update-6n5md"] Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.283271 4818 generic.go:334] "Generic (PLEG): container finished" podID="be50273d-b600-4461-8c58-69cf20892e97" containerID="0c5e9f19024b4862a8edc9390183eeca4ae34947a509ccd1a779ed5ade470b44" exitCode=0 Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.283369 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" event={"ID":"be50273d-b600-4461-8c58-69cf20892e97","Type":"ContainerDied","Data":"0c5e9f19024b4862a8edc9390183eeca4ae34947a509ccd1a779ed5ade470b44"} Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.286045 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-d25tm" event={"ID":"0b5ecd23-fa73-444e-a79c-166c4a25dff2","Type":"ContainerDied","Data":"c7bf2af9108106e72737e9c78d1d19076606c43f7efb1b68b8c57135c1f8154e"} Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.286077 4818 scope.go:117] "RemoveContainer" containerID="871491b91338481cc0c5af41a8ce042e475ec1cc84c51ebacd4f01c47b63edc5" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.286237 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-d25tm" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.295181 4818 generic.go:334] "Generic (PLEG): container finished" podID="6bf0813f-9f49-43b7-9e9a-8712e6a86159" containerID="0e35790c4bdc752e3a756f6b49e3294817e3332562c1a893fdf5ca8bf928fbed" exitCode=0 Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.295303 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-8p9jz" event={"ID":"6bf0813f-9f49-43b7-9e9a-8712e6a86159","Type":"ContainerDied","Data":"0e35790c4bdc752e3a756f6b49e3294817e3332562c1a893fdf5ca8bf928fbed"} Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.310345 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-srrh4" event={"ID":"997e3f50-bc25-4436-b03d-5e98770df25f","Type":"ContainerStarted","Data":"f0c950261f6a081d904eb5fd28ec64deea76cdfa8b3d903accf94962c15c3664"} Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.338948 4818 scope.go:117] "RemoveContainer" containerID="7d190cb5d11bb6790379c1b718a23058cf37d9a0cfc847c11f0221a59bed25d7" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.356665 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llzgr\" (UniqueName: \"kubernetes.io/projected/391197f9-7719-4fae-8698-ea3fe74bc4c0-kube-api-access-llzgr\") pod \"placement-9db7-account-create-update-6n5md\" (UID: \"391197f9-7719-4fae-8698-ea3fe74bc4c0\") " pod="openstack/placement-9db7-account-create-update-6n5md" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.356762 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9bb6d069-8823-4c5c-be08-d9bbb831e977-operator-scripts\") pod \"placement-db-create-md6rj\" (UID: \"9bb6d069-8823-4c5c-be08-d9bbb831e977\") " pod="openstack/placement-db-create-md6rj" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.356968 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg777\" (UniqueName: \"kubernetes.io/projected/9bb6d069-8823-4c5c-be08-d9bbb831e977-kube-api-access-vg777\") pod \"placement-db-create-md6rj\" (UID: \"9bb6d069-8823-4c5c-be08-d9bbb831e977\") " pod="openstack/placement-db-create-md6rj" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.357057 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/391197f9-7719-4fae-8698-ea3fe74bc4c0-operator-scripts\") pod \"placement-9db7-account-create-update-6n5md\" (UID: \"391197f9-7719-4fae-8698-ea3fe74bc4c0\") " pod="openstack/placement-9db7-account-create-update-6n5md" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.360598 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9bb6d069-8823-4c5c-be08-d9bbb831e977-operator-scripts\") pod \"placement-db-create-md6rj\" (UID: \"9bb6d069-8823-4c5c-be08-d9bbb831e977\") " pod="openstack/placement-db-create-md6rj" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.376627 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-srrh4" podStartSLOduration=8.376603677 podStartE2EDuration="8.376603677s" podCreationTimestamp="2025-12-03 06:44:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:44:59.354280975 +0000 UTC m=+1057.045889737" watchObservedRunningTime="2025-12-03 06:44:59.376603677 +0000 UTC m=+1057.068212429" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.390696 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg777\" (UniqueName: \"kubernetes.io/projected/9bb6d069-8823-4c5c-be08-d9bbb831e977-kube-api-access-vg777\") pod \"placement-db-create-md6rj\" (UID: \"9bb6d069-8823-4c5c-be08-d9bbb831e977\") " pod="openstack/placement-db-create-md6rj" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.435666 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-d25tm"] Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.444045 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-d25tm"] Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.458878 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-jfgjx"] Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.460652 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jfgjx" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.461608 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/391197f9-7719-4fae-8698-ea3fe74bc4c0-operator-scripts\") pod \"placement-9db7-account-create-update-6n5md\" (UID: \"391197f9-7719-4fae-8698-ea3fe74bc4c0\") " pod="openstack/placement-9db7-account-create-update-6n5md" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.461716 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llzgr\" (UniqueName: \"kubernetes.io/projected/391197f9-7719-4fae-8698-ea3fe74bc4c0-kube-api-access-llzgr\") pod \"placement-9db7-account-create-update-6n5md\" (UID: \"391197f9-7719-4fae-8698-ea3fe74bc4c0\") " pod="openstack/placement-9db7-account-create-update-6n5md" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.462639 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/391197f9-7719-4fae-8698-ea3fe74bc4c0-operator-scripts\") pod \"placement-9db7-account-create-update-6n5md\" (UID: \"391197f9-7719-4fae-8698-ea3fe74bc4c0\") " pod="openstack/placement-9db7-account-create-update-6n5md" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.462666 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-md6rj" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.465263 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-jfgjx"] Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.490568 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llzgr\" (UniqueName: \"kubernetes.io/projected/391197f9-7719-4fae-8698-ea3fe74bc4c0-kube-api-access-llzgr\") pod \"placement-9db7-account-create-update-6n5md\" (UID: \"391197f9-7719-4fae-8698-ea3fe74bc4c0\") " pod="openstack/placement-9db7-account-create-update-6n5md" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.551827 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-b465-account-create-update-2bd7d"] Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.553168 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b465-account-create-update-2bd7d" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.555684 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.558545 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b465-account-create-update-2bd7d"] Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.568090 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fktdl\" (UniqueName: \"kubernetes.io/projected/68aa86ae-c846-4321-aac9-b37e8837356f-kube-api-access-fktdl\") pod \"glance-db-create-jfgjx\" (UID: \"68aa86ae-c846-4321-aac9-b37e8837356f\") " pod="openstack/glance-db-create-jfgjx" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.568187 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68aa86ae-c846-4321-aac9-b37e8837356f-operator-scripts\") pod \"glance-db-create-jfgjx\" (UID: \"68aa86ae-c846-4321-aac9-b37e8837356f\") " pod="openstack/glance-db-create-jfgjx" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.568391 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9db7-account-create-update-6n5md" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.670646 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a7f2204-d32f-4497-b766-b8cbb659bb89-operator-scripts\") pod \"glance-b465-account-create-update-2bd7d\" (UID: \"5a7f2204-d32f-4497-b766-b8cbb659bb89\") " pod="openstack/glance-b465-account-create-update-2bd7d" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.671250 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fktdl\" (UniqueName: \"kubernetes.io/projected/68aa86ae-c846-4321-aac9-b37e8837356f-kube-api-access-fktdl\") pod \"glance-db-create-jfgjx\" (UID: \"68aa86ae-c846-4321-aac9-b37e8837356f\") " pod="openstack/glance-db-create-jfgjx" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.671341 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68aa86ae-c846-4321-aac9-b37e8837356f-operator-scripts\") pod \"glance-db-create-jfgjx\" (UID: \"68aa86ae-c846-4321-aac9-b37e8837356f\") " pod="openstack/glance-db-create-jfgjx" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.671392 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd677\" (UniqueName: \"kubernetes.io/projected/5a7f2204-d32f-4497-b766-b8cbb659bb89-kube-api-access-vd677\") pod \"glance-b465-account-create-update-2bd7d\" (UID: \"5a7f2204-d32f-4497-b766-b8cbb659bb89\") " pod="openstack/glance-b465-account-create-update-2bd7d" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.672328 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68aa86ae-c846-4321-aac9-b37e8837356f-operator-scripts\") pod \"glance-db-create-jfgjx\" (UID: \"68aa86ae-c846-4321-aac9-b37e8837356f\") " pod="openstack/glance-db-create-jfgjx" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.689500 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fktdl\" (UniqueName: \"kubernetes.io/projected/68aa86ae-c846-4321-aac9-b37e8837356f-kube-api-access-fktdl\") pod \"glance-db-create-jfgjx\" (UID: \"68aa86ae-c846-4321-aac9-b37e8837356f\") " pod="openstack/glance-db-create-jfgjx" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.780712 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd677\" (UniqueName: \"kubernetes.io/projected/5a7f2204-d32f-4497-b766-b8cbb659bb89-kube-api-access-vd677\") pod \"glance-b465-account-create-update-2bd7d\" (UID: \"5a7f2204-d32f-4497-b766-b8cbb659bb89\") " pod="openstack/glance-b465-account-create-update-2bd7d" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.780964 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a7f2204-d32f-4497-b766-b8cbb659bb89-operator-scripts\") pod \"glance-b465-account-create-update-2bd7d\" (UID: \"5a7f2204-d32f-4497-b766-b8cbb659bb89\") " pod="openstack/glance-b465-account-create-update-2bd7d" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.781893 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a7f2204-d32f-4497-b766-b8cbb659bb89-operator-scripts\") pod \"glance-b465-account-create-update-2bd7d\" (UID: \"5a7f2204-d32f-4497-b766-b8cbb659bb89\") " pod="openstack/glance-b465-account-create-update-2bd7d" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.796197 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd677\" (UniqueName: \"kubernetes.io/projected/5a7f2204-d32f-4497-b766-b8cbb659bb89-kube-api-access-vd677\") pod \"glance-b465-account-create-update-2bd7d\" (UID: \"5a7f2204-d32f-4497-b766-b8cbb659bb89\") " pod="openstack/glance-b465-account-create-update-2bd7d" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.842559 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jfgjx" Dec 03 06:44:59 crc kubenswrapper[4818]: I1203 06:44:59.879889 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b465-account-create-update-2bd7d" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.076808 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-15a6-account-create-update-rvmpz"] Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.093835 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-x7rhg"] Dec 03 06:45:00 crc kubenswrapper[4818]: W1203 06:45:00.106307 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod619d6969_7ea3_4680_ba99_061f79fcadcc.slice/crio-ff0e9540dc46d9e13aeb37033936c142b1180ca9bbe861494614fe468d42a9f0 WatchSource:0}: Error finding container ff0e9540dc46d9e13aeb37033936c142b1180ca9bbe861494614fe468d42a9f0: Status 404 returned error can't find the container with id ff0e9540dc46d9e13aeb37033936c142b1180ca9bbe861494614fe468d42a9f0 Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.226545 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv"] Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.233763 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.246024 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.247324 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.252868 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-md6rj"] Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.278799 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv"] Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.297950 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d7105fef-2722-4e9d-87ec-2553fbda781c-secret-volume\") pod \"collect-profiles-29412405-4cctv\" (UID: \"d7105fef-2722-4e9d-87ec-2553fbda781c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.298124 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vqh7\" (UniqueName: \"kubernetes.io/projected/d7105fef-2722-4e9d-87ec-2553fbda781c-kube-api-access-6vqh7\") pod \"collect-profiles-29412405-4cctv\" (UID: \"d7105fef-2722-4e9d-87ec-2553fbda781c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.300386 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d7105fef-2722-4e9d-87ec-2553fbda781c-config-volume\") pod \"collect-profiles-29412405-4cctv\" (UID: \"d7105fef-2722-4e9d-87ec-2553fbda781c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.319108 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-15a6-account-create-update-rvmpz" event={"ID":"619d6969-7ea3-4680-ba99-061f79fcadcc","Type":"ContainerStarted","Data":"ff0e9540dc46d9e13aeb37033936c142b1180ca9bbe861494614fe468d42a9f0"} Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.337163 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" event={"ID":"be50273d-b600-4461-8c58-69cf20892e97","Type":"ContainerStarted","Data":"3671afc10a586a757db89f4440fb2daeb881c236a2b0c217deb3500a9ebf6f3d"} Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.337251 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.340451 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-md6rj" event={"ID":"9bb6d069-8823-4c5c-be08-d9bbb831e977","Type":"ContainerStarted","Data":"f4c17f0d2ea6bdce7102e7b0ca57c9b95fb303d12dd836209abcf6bae92060fe"} Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.345392 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8df18ace-e467-4e73-9304-3ea16df510e8","Type":"ContainerStarted","Data":"76b766ec1fcdf5c1d9e4ecf00dbc988f4d84d840b2397aa8a27450216f8d09ad"} Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.345436 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8df18ace-e467-4e73-9304-3ea16df510e8","Type":"ContainerStarted","Data":"e7ebbbe73bb6921758152f8873b3d4f835d157d6e795382d305512367d0e524a"} Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.345554 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.346965 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-8p9jz" event={"ID":"6bf0813f-9f49-43b7-9e9a-8712e6a86159","Type":"ContainerStarted","Data":"42b5eb85bf09ae4c51edd5147653b5843bc2cf0821155b35637d68d685905682"} Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.347094 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.347375 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.369418 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-x7rhg" event={"ID":"19cc3001-7257-4267-a3e2-822ebf77b969","Type":"ContainerStarted","Data":"7a5aadecfdb080de06b852b269785d40c5df698e068bddaa491991e172ca0858"} Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.369743 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" podStartSLOduration=9.369729284 podStartE2EDuration="9.369729284s" podCreationTimestamp="2025-12-03 06:44:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:45:00.361350217 +0000 UTC m=+1058.052958969" watchObservedRunningTime="2025-12-03 06:45:00.369729284 +0000 UTC m=+1058.061338046" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.407929 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d7105fef-2722-4e9d-87ec-2553fbda781c-config-volume\") pod \"collect-profiles-29412405-4cctv\" (UID: \"d7105fef-2722-4e9d-87ec-2553fbda781c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.409897 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d7105fef-2722-4e9d-87ec-2553fbda781c-config-volume\") pod \"collect-profiles-29412405-4cctv\" (UID: \"d7105fef-2722-4e9d-87ec-2553fbda781c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.410184 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.251674971 podStartE2EDuration="9.410164434s" podCreationTimestamp="2025-12-03 06:44:51 +0000 UTC" firstStartedPulling="2025-12-03 06:44:53.118666995 +0000 UTC m=+1050.810275747" lastFinishedPulling="2025-12-03 06:44:59.277156468 +0000 UTC m=+1056.968765210" observedRunningTime="2025-12-03 06:45:00.40754644 +0000 UTC m=+1058.099155192" watchObservedRunningTime="2025-12-03 06:45:00.410164434 +0000 UTC m=+1058.101773186" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.414557 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d7105fef-2722-4e9d-87ec-2553fbda781c-secret-volume\") pod \"collect-profiles-29412405-4cctv\" (UID: \"d7105fef-2722-4e9d-87ec-2553fbda781c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.414776 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vqh7\" (UniqueName: \"kubernetes.io/projected/d7105fef-2722-4e9d-87ec-2553fbda781c-kube-api-access-6vqh7\") pod \"collect-profiles-29412405-4cctv\" (UID: \"d7105fef-2722-4e9d-87ec-2553fbda781c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.422485 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d7105fef-2722-4e9d-87ec-2553fbda781c-secret-volume\") pod \"collect-profiles-29412405-4cctv\" (UID: \"d7105fef-2722-4e9d-87ec-2553fbda781c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.431421 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-8p9jz" podStartSLOduration=9.43140458 podStartE2EDuration="9.43140458s" podCreationTimestamp="2025-12-03 06:44:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:45:00.430377304 +0000 UTC m=+1058.121986046" watchObservedRunningTime="2025-12-03 06:45:00.43140458 +0000 UTC m=+1058.123013332" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.436309 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vqh7\" (UniqueName: \"kubernetes.io/projected/d7105fef-2722-4e9d-87ec-2553fbda781c-kube-api-access-6vqh7\") pod \"collect-profiles-29412405-4cctv\" (UID: \"d7105fef-2722-4e9d-87ec-2553fbda781c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.479286 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.523433 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:45:00 crc kubenswrapper[4818]: E1203 06:45:00.523619 4818 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 06:45:00 crc kubenswrapper[4818]: E1203 06:45:00.523633 4818 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 06:45:00 crc kubenswrapper[4818]: E1203 06:45:00.523676 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift podName:520a6b29-d2d4-46e4-a5ed-989f2fa1df1f nodeName:}" failed. No retries permitted until 2025-12-03 06:45:08.523661531 +0000 UTC m=+1066.215270283 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift") pod "swift-storage-0" (UID: "520a6b29-d2d4-46e4-a5ed-989f2fa1df1f") : configmap "swift-ring-files" not found Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.538892 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-jfgjx"] Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.575379 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv" Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.651507 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b465-account-create-update-2bd7d"] Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.662026 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9db7-account-create-update-6n5md"] Dec 03 06:45:00 crc kubenswrapper[4818]: I1203 06:45:00.765034 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b5ecd23-fa73-444e-a79c-166c4a25dff2" path="/var/lib/kubelet/pods/0b5ecd23-fa73-444e-a79c-166c4a25dff2/volumes" Dec 03 06:45:01 crc kubenswrapper[4818]: W1203 06:45:01.300530 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod391197f9_7719_4fae_8698_ea3fe74bc4c0.slice/crio-71f6e3380375e4d75cbae80c4ae01d5689902c0262af71e0700a33073c1168f8 WatchSource:0}: Error finding container 71f6e3380375e4d75cbae80c4ae01d5689902c0262af71e0700a33073c1168f8: Status 404 returned error can't find the container with id 71f6e3380375e4d75cbae80c4ae01d5689902c0262af71e0700a33073c1168f8 Dec 03 06:45:01 crc kubenswrapper[4818]: W1203 06:45:01.311356 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68aa86ae_c846_4321_aac9_b37e8837356f.slice/crio-d681f556506e0c0b33e88ad92c3553a47135d56509f743e2ddb2ade32265006f WatchSource:0}: Error finding container d681f556506e0c0b33e88ad92c3553a47135d56509f743e2ddb2ade32265006f: Status 404 returned error can't find the container with id d681f556506e0c0b33e88ad92c3553a47135d56509f743e2ddb2ade32265006f Dec 03 06:45:01 crc kubenswrapper[4818]: I1203 06:45:01.377221 4818 generic.go:334] "Generic (PLEG): container finished" podID="619d6969-7ea3-4680-ba99-061f79fcadcc" containerID="912cef07c09a6dfc028f919c7b5c63422ad0d9bc0ef5e57df42e8a54d4d56729" exitCode=0 Dec 03 06:45:01 crc kubenswrapper[4818]: I1203 06:45:01.377282 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-15a6-account-create-update-rvmpz" event={"ID":"619d6969-7ea3-4680-ba99-061f79fcadcc","Type":"ContainerDied","Data":"912cef07c09a6dfc028f919c7b5c63422ad0d9bc0ef5e57df42e8a54d4d56729"} Dec 03 06:45:01 crc kubenswrapper[4818]: I1203 06:45:01.378472 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jfgjx" event={"ID":"68aa86ae-c846-4321-aac9-b37e8837356f","Type":"ContainerStarted","Data":"d681f556506e0c0b33e88ad92c3553a47135d56509f743e2ddb2ade32265006f"} Dec 03 06:45:01 crc kubenswrapper[4818]: I1203 06:45:01.379575 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9db7-account-create-update-6n5md" event={"ID":"391197f9-7719-4fae-8698-ea3fe74bc4c0","Type":"ContainerStarted","Data":"71f6e3380375e4d75cbae80c4ae01d5689902c0262af71e0700a33073c1168f8"} Dec 03 06:45:01 crc kubenswrapper[4818]: I1203 06:45:01.381462 4818 generic.go:334] "Generic (PLEG): container finished" podID="9bb6d069-8823-4c5c-be08-d9bbb831e977" containerID="27ead5511efe3d07eb90db2b555cacdba88c01daa96a0a979c13266a88ec897d" exitCode=0 Dec 03 06:45:01 crc kubenswrapper[4818]: I1203 06:45:01.381528 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-md6rj" event={"ID":"9bb6d069-8823-4c5c-be08-d9bbb831e977","Type":"ContainerDied","Data":"27ead5511efe3d07eb90db2b555cacdba88c01daa96a0a979c13266a88ec897d"} Dec 03 06:45:01 crc kubenswrapper[4818]: I1203 06:45:01.383430 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b465-account-create-update-2bd7d" event={"ID":"5a7f2204-d32f-4497-b766-b8cbb659bb89","Type":"ContainerStarted","Data":"13da79cec3edb502d07d2fc99dd6c6312107044a415e28a57092a6de29288035"} Dec 03 06:45:01 crc kubenswrapper[4818]: I1203 06:45:01.384907 4818 generic.go:334] "Generic (PLEG): container finished" podID="19cc3001-7257-4267-a3e2-822ebf77b969" containerID="ce9f7385328c736f4ccfa03ca017e311f37ba4f7ac1f4d4eb3c2304ca9011461" exitCode=0 Dec 03 06:45:01 crc kubenswrapper[4818]: I1203 06:45:01.384972 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-x7rhg" event={"ID":"19cc3001-7257-4267-a3e2-822ebf77b969","Type":"ContainerDied","Data":"ce9f7385328c736f4ccfa03ca017e311f37ba4f7ac1f4d4eb3c2304ca9011461"} Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.179288 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-x7rhg" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.253482 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19cc3001-7257-4267-a3e2-822ebf77b969-operator-scripts\") pod \"19cc3001-7257-4267-a3e2-822ebf77b969\" (UID: \"19cc3001-7257-4267-a3e2-822ebf77b969\") " Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.253560 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dxww\" (UniqueName: \"kubernetes.io/projected/19cc3001-7257-4267-a3e2-822ebf77b969-kube-api-access-8dxww\") pod \"19cc3001-7257-4267-a3e2-822ebf77b969\" (UID: \"19cc3001-7257-4267-a3e2-822ebf77b969\") " Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.255404 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19cc3001-7257-4267-a3e2-822ebf77b969-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "19cc3001-7257-4267-a3e2-822ebf77b969" (UID: "19cc3001-7257-4267-a3e2-822ebf77b969"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.260753 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19cc3001-7257-4267-a3e2-822ebf77b969-kube-api-access-8dxww" (OuterVolumeSpecName: "kube-api-access-8dxww") pod "19cc3001-7257-4267-a3e2-822ebf77b969" (UID: "19cc3001-7257-4267-a3e2-822ebf77b969"). InnerVolumeSpecName "kube-api-access-8dxww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.316334 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-15a6-account-create-update-rvmpz" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.348379 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-md6rj" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.354750 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/619d6969-7ea3-4680-ba99-061f79fcadcc-operator-scripts\") pod \"619d6969-7ea3-4680-ba99-061f79fcadcc\" (UID: \"619d6969-7ea3-4680-ba99-061f79fcadcc\") " Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.354908 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vg777\" (UniqueName: \"kubernetes.io/projected/9bb6d069-8823-4c5c-be08-d9bbb831e977-kube-api-access-vg777\") pod \"9bb6d069-8823-4c5c-be08-d9bbb831e977\" (UID: \"9bb6d069-8823-4c5c-be08-d9bbb831e977\") " Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.355076 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sxtb\" (UniqueName: \"kubernetes.io/projected/619d6969-7ea3-4680-ba99-061f79fcadcc-kube-api-access-5sxtb\") pod \"619d6969-7ea3-4680-ba99-061f79fcadcc\" (UID: \"619d6969-7ea3-4680-ba99-061f79fcadcc\") " Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.355158 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9bb6d069-8823-4c5c-be08-d9bbb831e977-operator-scripts\") pod \"9bb6d069-8823-4c5c-be08-d9bbb831e977\" (UID: \"9bb6d069-8823-4c5c-be08-d9bbb831e977\") " Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.355631 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/619d6969-7ea3-4680-ba99-061f79fcadcc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "619d6969-7ea3-4680-ba99-061f79fcadcc" (UID: "619d6969-7ea3-4680-ba99-061f79fcadcc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.356172 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/619d6969-7ea3-4680-ba99-061f79fcadcc-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.356190 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19cc3001-7257-4267-a3e2-822ebf77b969-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.356199 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dxww\" (UniqueName: \"kubernetes.io/projected/19cc3001-7257-4267-a3e2-822ebf77b969-kube-api-access-8dxww\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.356245 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bb6d069-8823-4c5c-be08-d9bbb831e977-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9bb6d069-8823-4c5c-be08-d9bbb831e977" (UID: "9bb6d069-8823-4c5c-be08-d9bbb831e977"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.364834 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/619d6969-7ea3-4680-ba99-061f79fcadcc-kube-api-access-5sxtb" (OuterVolumeSpecName: "kube-api-access-5sxtb") pod "619d6969-7ea3-4680-ba99-061f79fcadcc" (UID: "619d6969-7ea3-4680-ba99-061f79fcadcc"). InnerVolumeSpecName "kube-api-access-5sxtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.365060 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bb6d069-8823-4c5c-be08-d9bbb831e977-kube-api-access-vg777" (OuterVolumeSpecName: "kube-api-access-vg777") pod "9bb6d069-8823-4c5c-be08-d9bbb831e977" (UID: "9bb6d069-8823-4c5c-be08-d9bbb831e977"). InnerVolumeSpecName "kube-api-access-vg777". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.403429 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-x7rhg" event={"ID":"19cc3001-7257-4267-a3e2-822ebf77b969","Type":"ContainerDied","Data":"7a5aadecfdb080de06b852b269785d40c5df698e068bddaa491991e172ca0858"} Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.403467 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a5aadecfdb080de06b852b269785d40c5df698e068bddaa491991e172ca0858" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.403499 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-x7rhg" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.405848 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-15a6-account-create-update-rvmpz" event={"ID":"619d6969-7ea3-4680-ba99-061f79fcadcc","Type":"ContainerDied","Data":"ff0e9540dc46d9e13aeb37033936c142b1180ca9bbe861494614fe468d42a9f0"} Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.405876 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff0e9540dc46d9e13aeb37033936c142b1180ca9bbe861494614fe468d42a9f0" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.405952 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-15a6-account-create-update-rvmpz" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.408876 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-md6rj" event={"ID":"9bb6d069-8823-4c5c-be08-d9bbb831e977","Type":"ContainerDied","Data":"f4c17f0d2ea6bdce7102e7b0ca57c9b95fb303d12dd836209abcf6bae92060fe"} Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.408895 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4c17f0d2ea6bdce7102e7b0ca57c9b95fb303d12dd836209abcf6bae92060fe" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.408938 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-md6rj" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.445780 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv"] Dec 03 06:45:03 crc kubenswrapper[4818]: W1203 06:45:03.447809 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7105fef_2722_4e9d_87ec_2553fbda781c.slice/crio-18673be303642ba615374d512b6e8d97ca8eeff1fa9d39c060265bf16d0b4ae7 WatchSource:0}: Error finding container 18673be303642ba615374d512b6e8d97ca8eeff1fa9d39c060265bf16d0b4ae7: Status 404 returned error can't find the container with id 18673be303642ba615374d512b6e8d97ca8eeff1fa9d39c060265bf16d0b4ae7 Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.457361 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sxtb\" (UniqueName: \"kubernetes.io/projected/619d6969-7ea3-4680-ba99-061f79fcadcc-kube-api-access-5sxtb\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.457388 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9bb6d069-8823-4c5c-be08-d9bbb831e977-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:03 crc kubenswrapper[4818]: I1203 06:45:03.457398 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vg777\" (UniqueName: \"kubernetes.io/projected/9bb6d069-8823-4c5c-be08-d9bbb831e977-kube-api-access-vg777\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:04 crc kubenswrapper[4818]: I1203 06:45:04.423498 4818 generic.go:334] "Generic (PLEG): container finished" podID="391197f9-7719-4fae-8698-ea3fe74bc4c0" containerID="46d79510d850b2af9a38c2776abbeee09cb043e9d83bc8b27b5b905ddcab7942" exitCode=0 Dec 03 06:45:04 crc kubenswrapper[4818]: I1203 06:45:04.423598 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9db7-account-create-update-6n5md" event={"ID":"391197f9-7719-4fae-8698-ea3fe74bc4c0","Type":"ContainerDied","Data":"46d79510d850b2af9a38c2776abbeee09cb043e9d83bc8b27b5b905ddcab7942"} Dec 03 06:45:04 crc kubenswrapper[4818]: I1203 06:45:04.426350 4818 generic.go:334] "Generic (PLEG): container finished" podID="5a7f2204-d32f-4497-b766-b8cbb659bb89" containerID="472f472f6356bdf8aa74bf409dbdc7121ad48d0c2d215e3463d88a85a4f8300f" exitCode=0 Dec 03 06:45:04 crc kubenswrapper[4818]: I1203 06:45:04.426431 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b465-account-create-update-2bd7d" event={"ID":"5a7f2204-d32f-4497-b766-b8cbb659bb89","Type":"ContainerDied","Data":"472f472f6356bdf8aa74bf409dbdc7121ad48d0c2d215e3463d88a85a4f8300f"} Dec 03 06:45:04 crc kubenswrapper[4818]: I1203 06:45:04.428758 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-pkxxx" event={"ID":"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f","Type":"ContainerStarted","Data":"6bb7e5524aba1b7c7b542b8176d927c7d5e50bf99ce2f449d29f48b665ffae04"} Dec 03 06:45:04 crc kubenswrapper[4818]: I1203 06:45:04.434757 4818 generic.go:334] "Generic (PLEG): container finished" podID="d7105fef-2722-4e9d-87ec-2553fbda781c" containerID="54145502f815f6579c84f12f671d0ddd2d4ce8b94b904d29f5178b57ea8a4da3" exitCode=0 Dec 03 06:45:04 crc kubenswrapper[4818]: I1203 06:45:04.434877 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv" event={"ID":"d7105fef-2722-4e9d-87ec-2553fbda781c","Type":"ContainerDied","Data":"54145502f815f6579c84f12f671d0ddd2d4ce8b94b904d29f5178b57ea8a4da3"} Dec 03 06:45:04 crc kubenswrapper[4818]: I1203 06:45:04.434919 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv" event={"ID":"d7105fef-2722-4e9d-87ec-2553fbda781c","Type":"ContainerStarted","Data":"18673be303642ba615374d512b6e8d97ca8eeff1fa9d39c060265bf16d0b4ae7"} Dec 03 06:45:04 crc kubenswrapper[4818]: I1203 06:45:04.441282 4818 generic.go:334] "Generic (PLEG): container finished" podID="68aa86ae-c846-4321-aac9-b37e8837356f" containerID="3a79ad1243ecb7ef8281897632b9168bdae57e2a032e2c042356467c6f110fd2" exitCode=0 Dec 03 06:45:04 crc kubenswrapper[4818]: I1203 06:45:04.441340 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jfgjx" event={"ID":"68aa86ae-c846-4321-aac9-b37e8837356f","Type":"ContainerDied","Data":"3a79ad1243ecb7ef8281897632b9168bdae57e2a032e2c042356467c6f110fd2"} Dec 03 06:45:04 crc kubenswrapper[4818]: I1203 06:45:04.472400 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-pkxxx" podStartSLOduration=2.948854071 podStartE2EDuration="8.472382533s" podCreationTimestamp="2025-12-03 06:44:56 +0000 UTC" firstStartedPulling="2025-12-03 06:44:57.536495846 +0000 UTC m=+1055.228104598" lastFinishedPulling="2025-12-03 06:45:03.060024298 +0000 UTC m=+1060.751633060" observedRunningTime="2025-12-03 06:45:04.468437865 +0000 UTC m=+1062.160046617" watchObservedRunningTime="2025-12-03 06:45:04.472382533 +0000 UTC m=+1062.163991285" Dec 03 06:45:05 crc kubenswrapper[4818]: I1203 06:45:05.846257 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jfgjx" Dec 03 06:45:05 crc kubenswrapper[4818]: I1203 06:45:05.921164 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68aa86ae-c846-4321-aac9-b37e8837356f-operator-scripts\") pod \"68aa86ae-c846-4321-aac9-b37e8837356f\" (UID: \"68aa86ae-c846-4321-aac9-b37e8837356f\") " Dec 03 06:45:05 crc kubenswrapper[4818]: I1203 06:45:05.921253 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fktdl\" (UniqueName: \"kubernetes.io/projected/68aa86ae-c846-4321-aac9-b37e8837356f-kube-api-access-fktdl\") pod \"68aa86ae-c846-4321-aac9-b37e8837356f\" (UID: \"68aa86ae-c846-4321-aac9-b37e8837356f\") " Dec 03 06:45:05 crc kubenswrapper[4818]: I1203 06:45:05.922832 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68aa86ae-c846-4321-aac9-b37e8837356f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "68aa86ae-c846-4321-aac9-b37e8837356f" (UID: "68aa86ae-c846-4321-aac9-b37e8837356f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:05 crc kubenswrapper[4818]: I1203 06:45:05.926698 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68aa86ae-c846-4321-aac9-b37e8837356f-kube-api-access-fktdl" (OuterVolumeSpecName: "kube-api-access-fktdl") pod "68aa86ae-c846-4321-aac9-b37e8837356f" (UID: "68aa86ae-c846-4321-aac9-b37e8837356f"). InnerVolumeSpecName "kube-api-access-fktdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:05 crc kubenswrapper[4818]: I1203 06:45:05.959030 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv" Dec 03 06:45:05 crc kubenswrapper[4818]: I1203 06:45:05.963966 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b465-account-create-update-2bd7d" Dec 03 06:45:05 crc kubenswrapper[4818]: I1203 06:45:05.970596 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9db7-account-create-update-6n5md" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.023044 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llzgr\" (UniqueName: \"kubernetes.io/projected/391197f9-7719-4fae-8698-ea3fe74bc4c0-kube-api-access-llzgr\") pod \"391197f9-7719-4fae-8698-ea3fe74bc4c0\" (UID: \"391197f9-7719-4fae-8698-ea3fe74bc4c0\") " Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.023087 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd677\" (UniqueName: \"kubernetes.io/projected/5a7f2204-d32f-4497-b766-b8cbb659bb89-kube-api-access-vd677\") pod \"5a7f2204-d32f-4497-b766-b8cbb659bb89\" (UID: \"5a7f2204-d32f-4497-b766-b8cbb659bb89\") " Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.023114 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d7105fef-2722-4e9d-87ec-2553fbda781c-config-volume\") pod \"d7105fef-2722-4e9d-87ec-2553fbda781c\" (UID: \"d7105fef-2722-4e9d-87ec-2553fbda781c\") " Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.023289 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/391197f9-7719-4fae-8698-ea3fe74bc4c0-operator-scripts\") pod \"391197f9-7719-4fae-8698-ea3fe74bc4c0\" (UID: \"391197f9-7719-4fae-8698-ea3fe74bc4c0\") " Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.023309 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vqh7\" (UniqueName: \"kubernetes.io/projected/d7105fef-2722-4e9d-87ec-2553fbda781c-kube-api-access-6vqh7\") pod \"d7105fef-2722-4e9d-87ec-2553fbda781c\" (UID: \"d7105fef-2722-4e9d-87ec-2553fbda781c\") " Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.023344 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d7105fef-2722-4e9d-87ec-2553fbda781c-secret-volume\") pod \"d7105fef-2722-4e9d-87ec-2553fbda781c\" (UID: \"d7105fef-2722-4e9d-87ec-2553fbda781c\") " Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.023425 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a7f2204-d32f-4497-b766-b8cbb659bb89-operator-scripts\") pod \"5a7f2204-d32f-4497-b766-b8cbb659bb89\" (UID: \"5a7f2204-d32f-4497-b766-b8cbb659bb89\") " Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.023803 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a7f2204-d32f-4497-b766-b8cbb659bb89-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5a7f2204-d32f-4497-b766-b8cbb659bb89" (UID: "5a7f2204-d32f-4497-b766-b8cbb659bb89"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.023869 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7105fef-2722-4e9d-87ec-2553fbda781c-config-volume" (OuterVolumeSpecName: "config-volume") pod "d7105fef-2722-4e9d-87ec-2553fbda781c" (UID: "d7105fef-2722-4e9d-87ec-2553fbda781c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.023969 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fktdl\" (UniqueName: \"kubernetes.io/projected/68aa86ae-c846-4321-aac9-b37e8837356f-kube-api-access-fktdl\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.023991 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a7f2204-d32f-4497-b766-b8cbb659bb89-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.024004 4818 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d7105fef-2722-4e9d-87ec-2553fbda781c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.024017 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/68aa86ae-c846-4321-aac9-b37e8837356f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.024252 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/391197f9-7719-4fae-8698-ea3fe74bc4c0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "391197f9-7719-4fae-8698-ea3fe74bc4c0" (UID: "391197f9-7719-4fae-8698-ea3fe74bc4c0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.027283 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/391197f9-7719-4fae-8698-ea3fe74bc4c0-kube-api-access-llzgr" (OuterVolumeSpecName: "kube-api-access-llzgr") pod "391197f9-7719-4fae-8698-ea3fe74bc4c0" (UID: "391197f9-7719-4fae-8698-ea3fe74bc4c0"). InnerVolumeSpecName "kube-api-access-llzgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.027926 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7105fef-2722-4e9d-87ec-2553fbda781c-kube-api-access-6vqh7" (OuterVolumeSpecName: "kube-api-access-6vqh7") pod "d7105fef-2722-4e9d-87ec-2553fbda781c" (UID: "d7105fef-2722-4e9d-87ec-2553fbda781c"). InnerVolumeSpecName "kube-api-access-6vqh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.028974 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a7f2204-d32f-4497-b766-b8cbb659bb89-kube-api-access-vd677" (OuterVolumeSpecName: "kube-api-access-vd677") pod "5a7f2204-d32f-4497-b766-b8cbb659bb89" (UID: "5a7f2204-d32f-4497-b766-b8cbb659bb89"). InnerVolumeSpecName "kube-api-access-vd677". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.029157 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7105fef-2722-4e9d-87ec-2553fbda781c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d7105fef-2722-4e9d-87ec-2553fbda781c" (UID: "d7105fef-2722-4e9d-87ec-2553fbda781c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.125063 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/391197f9-7719-4fae-8698-ea3fe74bc4c0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.125103 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vqh7\" (UniqueName: \"kubernetes.io/projected/d7105fef-2722-4e9d-87ec-2553fbda781c-kube-api-access-6vqh7\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.125118 4818 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d7105fef-2722-4e9d-87ec-2553fbda781c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.125131 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llzgr\" (UniqueName: \"kubernetes.io/projected/391197f9-7719-4fae-8698-ea3fe74bc4c0-kube-api-access-llzgr\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.125144 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd677\" (UniqueName: \"kubernetes.io/projected/5a7f2204-d32f-4497-b766-b8cbb659bb89-kube-api-access-vd677\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.462466 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.462891 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv" event={"ID":"d7105fef-2722-4e9d-87ec-2553fbda781c","Type":"ContainerDied","Data":"18673be303642ba615374d512b6e8d97ca8eeff1fa9d39c060265bf16d0b4ae7"} Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.462987 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18673be303642ba615374d512b6e8d97ca8eeff1fa9d39c060265bf16d0b4ae7" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.467541 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jfgjx" event={"ID":"68aa86ae-c846-4321-aac9-b37e8837356f","Type":"ContainerDied","Data":"d681f556506e0c0b33e88ad92c3553a47135d56509f743e2ddb2ade32265006f"} Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.467623 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d681f556506e0c0b33e88ad92c3553a47135d56509f743e2ddb2ade32265006f" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.467553 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jfgjx" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.471218 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9db7-account-create-update-6n5md" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.471290 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9db7-account-create-update-6n5md" event={"ID":"391197f9-7719-4fae-8698-ea3fe74bc4c0","Type":"ContainerDied","Data":"71f6e3380375e4d75cbae80c4ae01d5689902c0262af71e0700a33073c1168f8"} Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.471375 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71f6e3380375e4d75cbae80c4ae01d5689902c0262af71e0700a33073c1168f8" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.474292 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b465-account-create-update-2bd7d" event={"ID":"5a7f2204-d32f-4497-b766-b8cbb659bb89","Type":"ContainerDied","Data":"13da79cec3edb502d07d2fc99dd6c6312107044a415e28a57092a6de29288035"} Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.474341 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13da79cec3edb502d07d2fc99dd6c6312107044a415e28a57092a6de29288035" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.474401 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b465-account-create-update-2bd7d" Dec 03 06:45:06 crc kubenswrapper[4818]: I1203 06:45:06.879084 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" Dec 03 06:45:07 crc kubenswrapper[4818]: I1203 06:45:07.384960 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:45:07 crc kubenswrapper[4818]: I1203 06:45:07.441044 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-6sprz"] Dec 03 06:45:07 crc kubenswrapper[4818]: I1203 06:45:07.483997 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" podUID="be50273d-b600-4461-8c58-69cf20892e97" containerName="dnsmasq-dns" containerID="cri-o://3671afc10a586a757db89f4440fb2daeb881c236a2b0c217deb3500a9ebf6f3d" gracePeriod=10 Dec 03 06:45:07 crc kubenswrapper[4818]: I1203 06:45:07.979058 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.069468 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be50273d-b600-4461-8c58-69cf20892e97-dns-svc\") pod \"be50273d-b600-4461-8c58-69cf20892e97\" (UID: \"be50273d-b600-4461-8c58-69cf20892e97\") " Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.069710 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be50273d-b600-4461-8c58-69cf20892e97-config\") pod \"be50273d-b600-4461-8c58-69cf20892e97\" (UID: \"be50273d-b600-4461-8c58-69cf20892e97\") " Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.069761 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxm44\" (UniqueName: \"kubernetes.io/projected/be50273d-b600-4461-8c58-69cf20892e97-kube-api-access-pxm44\") pod \"be50273d-b600-4461-8c58-69cf20892e97\" (UID: \"be50273d-b600-4461-8c58-69cf20892e97\") " Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.085035 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be50273d-b600-4461-8c58-69cf20892e97-kube-api-access-pxm44" (OuterVolumeSpecName: "kube-api-access-pxm44") pod "be50273d-b600-4461-8c58-69cf20892e97" (UID: "be50273d-b600-4461-8c58-69cf20892e97"). InnerVolumeSpecName "kube-api-access-pxm44". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.111299 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be50273d-b600-4461-8c58-69cf20892e97-config" (OuterVolumeSpecName: "config") pod "be50273d-b600-4461-8c58-69cf20892e97" (UID: "be50273d-b600-4461-8c58-69cf20892e97"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.111596 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be50273d-b600-4461-8c58-69cf20892e97-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "be50273d-b600-4461-8c58-69cf20892e97" (UID: "be50273d-b600-4461-8c58-69cf20892e97"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.171443 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be50273d-b600-4461-8c58-69cf20892e97-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.171479 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxm44\" (UniqueName: \"kubernetes.io/projected/be50273d-b600-4461-8c58-69cf20892e97-kube-api-access-pxm44\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.171493 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/be50273d-b600-4461-8c58-69cf20892e97-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.495847 4818 generic.go:334] "Generic (PLEG): container finished" podID="be50273d-b600-4461-8c58-69cf20892e97" containerID="3671afc10a586a757db89f4440fb2daeb881c236a2b0c217deb3500a9ebf6f3d" exitCode=0 Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.495898 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.495906 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" event={"ID":"be50273d-b600-4461-8c58-69cf20892e97","Type":"ContainerDied","Data":"3671afc10a586a757db89f4440fb2daeb881c236a2b0c217deb3500a9ebf6f3d"} Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.495942 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-6sprz" event={"ID":"be50273d-b600-4461-8c58-69cf20892e97","Type":"ContainerDied","Data":"79e5b4a6e916af18aebb3e3cb1450c83d9b354e0df16104b21123d9c75cdc579"} Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.495968 4818 scope.go:117] "RemoveContainer" containerID="3671afc10a586a757db89f4440fb2daeb881c236a2b0c217deb3500a9ebf6f3d" Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.535749 4818 scope.go:117] "RemoveContainer" containerID="0c5e9f19024b4862a8edc9390183eeca4ae34947a509ccd1a779ed5ade470b44" Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.539450 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-6sprz"] Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.548900 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-6sprz"] Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.562473 4818 scope.go:117] "RemoveContainer" containerID="3671afc10a586a757db89f4440fb2daeb881c236a2b0c217deb3500a9ebf6f3d" Dec 03 06:45:08 crc kubenswrapper[4818]: E1203 06:45:08.566389 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3671afc10a586a757db89f4440fb2daeb881c236a2b0c217deb3500a9ebf6f3d\": container with ID starting with 3671afc10a586a757db89f4440fb2daeb881c236a2b0c217deb3500a9ebf6f3d not found: ID does not exist" containerID="3671afc10a586a757db89f4440fb2daeb881c236a2b0c217deb3500a9ebf6f3d" Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.566440 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3671afc10a586a757db89f4440fb2daeb881c236a2b0c217deb3500a9ebf6f3d"} err="failed to get container status \"3671afc10a586a757db89f4440fb2daeb881c236a2b0c217deb3500a9ebf6f3d\": rpc error: code = NotFound desc = could not find container \"3671afc10a586a757db89f4440fb2daeb881c236a2b0c217deb3500a9ebf6f3d\": container with ID starting with 3671afc10a586a757db89f4440fb2daeb881c236a2b0c217deb3500a9ebf6f3d not found: ID does not exist" Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.566471 4818 scope.go:117] "RemoveContainer" containerID="0c5e9f19024b4862a8edc9390183eeca4ae34947a509ccd1a779ed5ade470b44" Dec 03 06:45:08 crc kubenswrapper[4818]: E1203 06:45:08.566949 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c5e9f19024b4862a8edc9390183eeca4ae34947a509ccd1a779ed5ade470b44\": container with ID starting with 0c5e9f19024b4862a8edc9390183eeca4ae34947a509ccd1a779ed5ade470b44 not found: ID does not exist" containerID="0c5e9f19024b4862a8edc9390183eeca4ae34947a509ccd1a779ed5ade470b44" Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.566982 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c5e9f19024b4862a8edc9390183eeca4ae34947a509ccd1a779ed5ade470b44"} err="failed to get container status \"0c5e9f19024b4862a8edc9390183eeca4ae34947a509ccd1a779ed5ade470b44\": rpc error: code = NotFound desc = could not find container \"0c5e9f19024b4862a8edc9390183eeca4ae34947a509ccd1a779ed5ade470b44\": container with ID starting with 0c5e9f19024b4862a8edc9390183eeca4ae34947a509ccd1a779ed5ade470b44 not found: ID does not exist" Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.578043 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:45:08 crc kubenswrapper[4818]: E1203 06:45:08.578291 4818 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 06:45:08 crc kubenswrapper[4818]: E1203 06:45:08.578326 4818 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 06:45:08 crc kubenswrapper[4818]: E1203 06:45:08.578392 4818 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift podName:520a6b29-d2d4-46e4-a5ed-989f2fa1df1f nodeName:}" failed. No retries permitted until 2025-12-03 06:45:24.578370883 +0000 UTC m=+1082.269979635 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift") pod "swift-storage-0" (UID: "520a6b29-d2d4-46e4-a5ed-989f2fa1df1f") : configmap "swift-ring-files" not found Dec 03 06:45:08 crc kubenswrapper[4818]: I1203 06:45:08.755299 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be50273d-b600-4461-8c58-69cf20892e97" path="/var/lib/kubelet/pods/be50273d-b600-4461-8c58-69cf20892e97/volumes" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.697200 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-p57tz"] Dec 03 06:45:09 crc kubenswrapper[4818]: E1203 06:45:09.697515 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bb6d069-8823-4c5c-be08-d9bbb831e977" containerName="mariadb-database-create" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.697526 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bb6d069-8823-4c5c-be08-d9bbb831e977" containerName="mariadb-database-create" Dec 03 06:45:09 crc kubenswrapper[4818]: E1203 06:45:09.697536 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a7f2204-d32f-4497-b766-b8cbb659bb89" containerName="mariadb-account-create-update" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.697542 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a7f2204-d32f-4497-b766-b8cbb659bb89" containerName="mariadb-account-create-update" Dec 03 06:45:09 crc kubenswrapper[4818]: E1203 06:45:09.697555 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68aa86ae-c846-4321-aac9-b37e8837356f" containerName="mariadb-database-create" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.697561 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="68aa86ae-c846-4321-aac9-b37e8837356f" containerName="mariadb-database-create" Dec 03 06:45:09 crc kubenswrapper[4818]: E1203 06:45:09.697571 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="391197f9-7719-4fae-8698-ea3fe74bc4c0" containerName="mariadb-account-create-update" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.697576 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="391197f9-7719-4fae-8698-ea3fe74bc4c0" containerName="mariadb-account-create-update" Dec 03 06:45:09 crc kubenswrapper[4818]: E1203 06:45:09.697588 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7105fef-2722-4e9d-87ec-2553fbda781c" containerName="collect-profiles" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.697593 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7105fef-2722-4e9d-87ec-2553fbda781c" containerName="collect-profiles" Dec 03 06:45:09 crc kubenswrapper[4818]: E1203 06:45:09.697602 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19cc3001-7257-4267-a3e2-822ebf77b969" containerName="mariadb-database-create" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.697607 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="19cc3001-7257-4267-a3e2-822ebf77b969" containerName="mariadb-database-create" Dec 03 06:45:09 crc kubenswrapper[4818]: E1203 06:45:09.697619 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="619d6969-7ea3-4680-ba99-061f79fcadcc" containerName="mariadb-account-create-update" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.697625 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="619d6969-7ea3-4680-ba99-061f79fcadcc" containerName="mariadb-account-create-update" Dec 03 06:45:09 crc kubenswrapper[4818]: E1203 06:45:09.697641 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be50273d-b600-4461-8c58-69cf20892e97" containerName="init" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.697647 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="be50273d-b600-4461-8c58-69cf20892e97" containerName="init" Dec 03 06:45:09 crc kubenswrapper[4818]: E1203 06:45:09.697657 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be50273d-b600-4461-8c58-69cf20892e97" containerName="dnsmasq-dns" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.697662 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="be50273d-b600-4461-8c58-69cf20892e97" containerName="dnsmasq-dns" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.697869 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="be50273d-b600-4461-8c58-69cf20892e97" containerName="dnsmasq-dns" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.697902 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a7f2204-d32f-4497-b766-b8cbb659bb89" containerName="mariadb-account-create-update" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.697920 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="391197f9-7719-4fae-8698-ea3fe74bc4c0" containerName="mariadb-account-create-update" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.697942 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bb6d069-8823-4c5c-be08-d9bbb831e977" containerName="mariadb-database-create" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.697962 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="19cc3001-7257-4267-a3e2-822ebf77b969" containerName="mariadb-database-create" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.697990 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7105fef-2722-4e9d-87ec-2553fbda781c" containerName="collect-profiles" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.698005 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="619d6969-7ea3-4680-ba99-061f79fcadcc" containerName="mariadb-account-create-update" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.698025 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="68aa86ae-c846-4321-aac9-b37e8837356f" containerName="mariadb-database-create" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.698658 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-p57tz" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.703486 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-zqbs6" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.703486 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.763189 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-p57tz"] Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.804111 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-combined-ca-bundle\") pod \"glance-db-sync-p57tz\" (UID: \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\") " pod="openstack/glance-db-sync-p57tz" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.804183 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zld56\" (UniqueName: \"kubernetes.io/projected/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-kube-api-access-zld56\") pod \"glance-db-sync-p57tz\" (UID: \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\") " pod="openstack/glance-db-sync-p57tz" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.804236 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-config-data\") pod \"glance-db-sync-p57tz\" (UID: \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\") " pod="openstack/glance-db-sync-p57tz" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.804261 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-db-sync-config-data\") pod \"glance-db-sync-p57tz\" (UID: \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\") " pod="openstack/glance-db-sync-p57tz" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.905164 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-combined-ca-bundle\") pod \"glance-db-sync-p57tz\" (UID: \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\") " pod="openstack/glance-db-sync-p57tz" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.905227 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zld56\" (UniqueName: \"kubernetes.io/projected/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-kube-api-access-zld56\") pod \"glance-db-sync-p57tz\" (UID: \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\") " pod="openstack/glance-db-sync-p57tz" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.905253 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-config-data\") pod \"glance-db-sync-p57tz\" (UID: \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\") " pod="openstack/glance-db-sync-p57tz" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.905277 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-db-sync-config-data\") pod \"glance-db-sync-p57tz\" (UID: \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\") " pod="openstack/glance-db-sync-p57tz" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.911435 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-config-data\") pod \"glance-db-sync-p57tz\" (UID: \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\") " pod="openstack/glance-db-sync-p57tz" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.911936 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-combined-ca-bundle\") pod \"glance-db-sync-p57tz\" (UID: \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\") " pod="openstack/glance-db-sync-p57tz" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.913459 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-db-sync-config-data\") pod \"glance-db-sync-p57tz\" (UID: \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\") " pod="openstack/glance-db-sync-p57tz" Dec 03 06:45:09 crc kubenswrapper[4818]: I1203 06:45:09.922440 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zld56\" (UniqueName: \"kubernetes.io/projected/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-kube-api-access-zld56\") pod \"glance-db-sync-p57tz\" (UID: \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\") " pod="openstack/glance-db-sync-p57tz" Dec 03 06:45:10 crc kubenswrapper[4818]: I1203 06:45:10.064887 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-p57tz" Dec 03 06:45:10 crc kubenswrapper[4818]: I1203 06:45:10.515413 4818 generic.go:334] "Generic (PLEG): container finished" podID="337c7d6b-71fc-4044-98ec-de8bfbf5dd6f" containerID="6bb7e5524aba1b7c7b542b8176d927c7d5e50bf99ce2f449d29f48b665ffae04" exitCode=0 Dec 03 06:45:10 crc kubenswrapper[4818]: I1203 06:45:10.515502 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-pkxxx" event={"ID":"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f","Type":"ContainerDied","Data":"6bb7e5524aba1b7c7b542b8176d927c7d5e50bf99ce2f449d29f48b665ffae04"} Dec 03 06:45:10 crc kubenswrapper[4818]: I1203 06:45:10.598431 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-p57tz"] Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.526841 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-p57tz" event={"ID":"ba5ae97f-71e8-4dc1-bcd4-673e558c9693","Type":"ContainerStarted","Data":"a36f2ebc02e5ea46d3914af488b14b60e216658a93469404cb74da964c2379c9"} Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.828876 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.841901 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-dispersionconf\") pod \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.842013 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-ring-data-devices\") pod \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.842041 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-combined-ca-bundle\") pod \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.842067 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-scripts\") pod \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.842123 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wz5cg\" (UniqueName: \"kubernetes.io/projected/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-kube-api-access-wz5cg\") pod \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.843040 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "337c7d6b-71fc-4044-98ec-de8bfbf5dd6f" (UID: "337c7d6b-71fc-4044-98ec-de8bfbf5dd6f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.845961 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "337c7d6b-71fc-4044-98ec-de8bfbf5dd6f" (UID: "337c7d6b-71fc-4044-98ec-de8bfbf5dd6f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.846442 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-etc-swift\") pod \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.846520 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-swiftconf\") pod \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\" (UID: \"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f\") " Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.848516 4818 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.848543 4818 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.855269 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-kube-api-access-wz5cg" (OuterVolumeSpecName: "kube-api-access-wz5cg") pod "337c7d6b-71fc-4044-98ec-de8bfbf5dd6f" (UID: "337c7d6b-71fc-4044-98ec-de8bfbf5dd6f"). InnerVolumeSpecName "kube-api-access-wz5cg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.855857 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "337c7d6b-71fc-4044-98ec-de8bfbf5dd6f" (UID: "337c7d6b-71fc-4044-98ec-de8bfbf5dd6f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.872791 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "337c7d6b-71fc-4044-98ec-de8bfbf5dd6f" (UID: "337c7d6b-71fc-4044-98ec-de8bfbf5dd6f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.876420 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "337c7d6b-71fc-4044-98ec-de8bfbf5dd6f" (UID: "337c7d6b-71fc-4044-98ec-de8bfbf5dd6f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.885768 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-scripts" (OuterVolumeSpecName: "scripts") pod "337c7d6b-71fc-4044-98ec-de8bfbf5dd6f" (UID: "337c7d6b-71fc-4044-98ec-de8bfbf5dd6f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.950573 4818 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.950624 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.950641 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.950655 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wz5cg\" (UniqueName: \"kubernetes.io/projected/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-kube-api-access-wz5cg\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:11 crc kubenswrapper[4818]: I1203 06:45:11.950668 4818 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/337c7d6b-71fc-4044-98ec-de8bfbf5dd6f-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:12 crc kubenswrapper[4818]: I1203 06:45:12.393069 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 03 06:45:12 crc kubenswrapper[4818]: I1203 06:45:12.535990 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-pkxxx" event={"ID":"337c7d6b-71fc-4044-98ec-de8bfbf5dd6f","Type":"ContainerDied","Data":"576d6118befbfed8152e7db32e7ca031f9797b3b2c6dc18ebaeb97bcba5b89df"} Dec 03 06:45:12 crc kubenswrapper[4818]: I1203 06:45:12.536042 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="576d6118befbfed8152e7db32e7ca031f9797b3b2c6dc18ebaeb97bcba5b89df" Dec 03 06:45:12 crc kubenswrapper[4818]: I1203 06:45:12.536116 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pkxxx" Dec 03 06:45:15 crc kubenswrapper[4818]: I1203 06:45:15.470356 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-kvg5t" podUID="04bf523a-de6a-49a0-93ad-cca1e22939c6" containerName="ovn-controller" probeResult="failure" output=< Dec 03 06:45:15 crc kubenswrapper[4818]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 06:45:15 crc kubenswrapper[4818]: > Dec 03 06:45:15 crc kubenswrapper[4818]: I1203 06:45:15.482103 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:45:15 crc kubenswrapper[4818]: I1203 06:45:15.483660 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-brxh4" Dec 03 06:45:15 crc kubenswrapper[4818]: I1203 06:45:15.816242 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-kvg5t-config-qtwrm"] Dec 03 06:45:15 crc kubenswrapper[4818]: E1203 06:45:15.822597 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="337c7d6b-71fc-4044-98ec-de8bfbf5dd6f" containerName="swift-ring-rebalance" Dec 03 06:45:15 crc kubenswrapper[4818]: I1203 06:45:15.822639 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="337c7d6b-71fc-4044-98ec-de8bfbf5dd6f" containerName="swift-ring-rebalance" Dec 03 06:45:15 crc kubenswrapper[4818]: I1203 06:45:15.822897 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="337c7d6b-71fc-4044-98ec-de8bfbf5dd6f" containerName="swift-ring-rebalance" Dec 03 06:45:15 crc kubenswrapper[4818]: I1203 06:45:15.823590 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:15 crc kubenswrapper[4818]: I1203 06:45:15.828915 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 06:45:15 crc kubenswrapper[4818]: I1203 06:45:15.829951 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kvg5t-config-qtwrm"] Dec 03 06:45:15 crc kubenswrapper[4818]: I1203 06:45:15.968600 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-log-ovn\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:15 crc kubenswrapper[4818]: I1203 06:45:15.968686 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d74d063c-c0e6-401c-857d-e5529f2df21a-additional-scripts\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:15 crc kubenswrapper[4818]: I1203 06:45:15.968710 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxhdq\" (UniqueName: \"kubernetes.io/projected/d74d063c-c0e6-401c-857d-e5529f2df21a-kube-api-access-jxhdq\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:15 crc kubenswrapper[4818]: I1203 06:45:15.968855 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d74d063c-c0e6-401c-857d-e5529f2df21a-scripts\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:15 crc kubenswrapper[4818]: I1203 06:45:15.968894 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-run-ovn\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:15 crc kubenswrapper[4818]: I1203 06:45:15.968935 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-run\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:16 crc kubenswrapper[4818]: I1203 06:45:16.070562 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-run\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:16 crc kubenswrapper[4818]: I1203 06:45:16.070692 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-log-ovn\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:16 crc kubenswrapper[4818]: I1203 06:45:16.070730 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d74d063c-c0e6-401c-857d-e5529f2df21a-additional-scripts\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:16 crc kubenswrapper[4818]: I1203 06:45:16.070756 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxhdq\" (UniqueName: \"kubernetes.io/projected/d74d063c-c0e6-401c-857d-e5529f2df21a-kube-api-access-jxhdq\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:16 crc kubenswrapper[4818]: I1203 06:45:16.070806 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d74d063c-c0e6-401c-857d-e5529f2df21a-scripts\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:16 crc kubenswrapper[4818]: I1203 06:45:16.070851 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-run-ovn\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:16 crc kubenswrapper[4818]: I1203 06:45:16.070990 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-run-ovn\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:16 crc kubenswrapper[4818]: I1203 06:45:16.070992 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-run\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:16 crc kubenswrapper[4818]: I1203 06:45:16.071023 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-log-ovn\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:16 crc kubenswrapper[4818]: I1203 06:45:16.072216 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d74d063c-c0e6-401c-857d-e5529f2df21a-additional-scripts\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:16 crc kubenswrapper[4818]: I1203 06:45:16.073361 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d74d063c-c0e6-401c-857d-e5529f2df21a-scripts\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:16 crc kubenswrapper[4818]: I1203 06:45:16.133290 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxhdq\" (UniqueName: \"kubernetes.io/projected/d74d063c-c0e6-401c-857d-e5529f2df21a-kube-api-access-jxhdq\") pod \"ovn-controller-kvg5t-config-qtwrm\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:16 crc kubenswrapper[4818]: I1203 06:45:16.222516 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:20 crc kubenswrapper[4818]: I1203 06:45:20.474303 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-kvg5t" podUID="04bf523a-de6a-49a0-93ad-cca1e22939c6" containerName="ovn-controller" probeResult="failure" output=< Dec 03 06:45:20 crc kubenswrapper[4818]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 06:45:20 crc kubenswrapper[4818]: > Dec 03 06:45:22 crc kubenswrapper[4818]: I1203 06:45:22.632578 4818 generic.go:334] "Generic (PLEG): container finished" podID="df9ab7c7-152b-4938-9ef5-d0588111b8a2" containerID="651b7927b0f2dd739e06ae0b1b47c6bf30fb2fa4bce7e634b77e016a5d7e4fce" exitCode=0 Dec 03 06:45:22 crc kubenswrapper[4818]: I1203 06:45:22.632663 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"df9ab7c7-152b-4938-9ef5-d0588111b8a2","Type":"ContainerDied","Data":"651b7927b0f2dd739e06ae0b1b47c6bf30fb2fa4bce7e634b77e016a5d7e4fce"} Dec 03 06:45:23 crc kubenswrapper[4818]: I1203 06:45:23.643152 4818 generic.go:334] "Generic (PLEG): container finished" podID="0e751691-c660-4527-a9a6-61505f83c6ff" containerID="303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0" exitCode=0 Dec 03 06:45:23 crc kubenswrapper[4818]: I1203 06:45:23.643241 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0e751691-c660-4527-a9a6-61505f83c6ff","Type":"ContainerDied","Data":"303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0"} Dec 03 06:45:24 crc kubenswrapper[4818]: I1203 06:45:24.616018 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:45:24 crc kubenswrapper[4818]: I1203 06:45:24.630513 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/520a6b29-d2d4-46e4-a5ed-989f2fa1df1f-etc-swift\") pod \"swift-storage-0\" (UID: \"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f\") " pod="openstack/swift-storage-0" Dec 03 06:45:24 crc kubenswrapper[4818]: I1203 06:45:24.905801 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 06:45:25 crc kubenswrapper[4818]: E1203 06:45:25.343064 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 03 06:45:25 crc kubenswrapper[4818]: E1203 06:45:25.344108 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zld56,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-p57tz_openstack(ba5ae97f-71e8-4dc1-bcd4-673e558c9693): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:45:25 crc kubenswrapper[4818]: E1203 06:45:25.345756 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-p57tz" podUID="ba5ae97f-71e8-4dc1-bcd4-673e558c9693" Dec 03 06:45:25 crc kubenswrapper[4818]: I1203 06:45:25.472598 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-kvg5t" podUID="04bf523a-de6a-49a0-93ad-cca1e22939c6" containerName="ovn-controller" probeResult="failure" output=< Dec 03 06:45:25 crc kubenswrapper[4818]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 06:45:25 crc kubenswrapper[4818]: > Dec 03 06:45:25 crc kubenswrapper[4818]: I1203 06:45:25.660085 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"df9ab7c7-152b-4938-9ef5-d0588111b8a2","Type":"ContainerStarted","Data":"b14a6d3062e8c0e207497d3cecb33774c99ddc3ee3d60372949002a157583eed"} Dec 03 06:45:25 crc kubenswrapper[4818]: I1203 06:45:25.660453 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:45:25 crc kubenswrapper[4818]: E1203 06:45:25.661397 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-p57tz" podUID="ba5ae97f-71e8-4dc1-bcd4-673e558c9693" Dec 03 06:45:25 crc kubenswrapper[4818]: I1203 06:45:25.682364 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=39.797967434 podStartE2EDuration="1m11.68234728s" podCreationTimestamp="2025-12-03 06:44:14 +0000 UTC" firstStartedPulling="2025-12-03 06:44:16.853341189 +0000 UTC m=+1014.544949941" lastFinishedPulling="2025-12-03 06:44:48.737721045 +0000 UTC m=+1046.429329787" observedRunningTime="2025-12-03 06:45:25.677533681 +0000 UTC m=+1083.369142433" watchObservedRunningTime="2025-12-03 06:45:25.68234728 +0000 UTC m=+1083.373956032" Dec 03 06:45:25 crc kubenswrapper[4818]: I1203 06:45:25.810278 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kvg5t-config-qtwrm"] Dec 03 06:45:25 crc kubenswrapper[4818]: I1203 06:45:25.939957 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 06:45:25 crc kubenswrapper[4818]: W1203 06:45:25.957786 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod520a6b29_d2d4_46e4_a5ed_989f2fa1df1f.slice/crio-c73dd060b8629973fadbd158ce5ab548c3ec9e60fc0a368a98a46168f60c5e9f WatchSource:0}: Error finding container c73dd060b8629973fadbd158ce5ab548c3ec9e60fc0a368a98a46168f60c5e9f: Status 404 returned error can't find the container with id c73dd060b8629973fadbd158ce5ab548c3ec9e60fc0a368a98a46168f60c5e9f Dec 03 06:45:26 crc kubenswrapper[4818]: I1203 06:45:26.667672 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f","Type":"ContainerStarted","Data":"c73dd060b8629973fadbd158ce5ab548c3ec9e60fc0a368a98a46168f60c5e9f"} Dec 03 06:45:26 crc kubenswrapper[4818]: I1203 06:45:26.670435 4818 generic.go:334] "Generic (PLEG): container finished" podID="d74d063c-c0e6-401c-857d-e5529f2df21a" containerID="3ac3909bb7ac5d22f52702b363772b4b063da74e54668549e4d542916d951527" exitCode=0 Dec 03 06:45:26 crc kubenswrapper[4818]: I1203 06:45:26.670500 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kvg5t-config-qtwrm" event={"ID":"d74d063c-c0e6-401c-857d-e5529f2df21a","Type":"ContainerDied","Data":"3ac3909bb7ac5d22f52702b363772b4b063da74e54668549e4d542916d951527"} Dec 03 06:45:26 crc kubenswrapper[4818]: I1203 06:45:26.670563 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kvg5t-config-qtwrm" event={"ID":"d74d063c-c0e6-401c-857d-e5529f2df21a","Type":"ContainerStarted","Data":"8697403d833dd9d3174cc79098e7da3ce015633580a91a756cbd9de80656eb59"} Dec 03 06:45:26 crc kubenswrapper[4818]: I1203 06:45:26.673438 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0e751691-c660-4527-a9a6-61505f83c6ff","Type":"ContainerStarted","Data":"5b437a91614dddcbffaba1e3457305c6be3910b1ccccef8d5c8a44875396d9f1"} Dec 03 06:45:26 crc kubenswrapper[4818]: I1203 06:45:26.717380 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=40.687024298 podStartE2EDuration="1m12.717348934s" podCreationTimestamp="2025-12-03 06:44:14 +0000 UTC" firstStartedPulling="2025-12-03 06:44:16.707685177 +0000 UTC m=+1014.399293929" lastFinishedPulling="2025-12-03 06:44:48.738009823 +0000 UTC m=+1046.429618565" observedRunningTime="2025-12-03 06:45:26.712360601 +0000 UTC m=+1084.403969363" watchObservedRunningTime="2025-12-03 06:45:26.717348934 +0000 UTC m=+1084.408957686" Dec 03 06:45:27 crc kubenswrapper[4818]: I1203 06:45:27.683218 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f","Type":"ContainerStarted","Data":"c719a0160f2ebe614b4b017bcd377c789b7eb0c46d68dc21357e9b288f8dc95d"} Dec 03 06:45:27 crc kubenswrapper[4818]: I1203 06:45:27.683284 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f","Type":"ContainerStarted","Data":"f8dcffd1b72d3ce9323cf4332482fd92055fdfa8afa0f597c8c36654a92c7499"} Dec 03 06:45:27 crc kubenswrapper[4818]: I1203 06:45:27.683300 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f","Type":"ContainerStarted","Data":"fe53b6a049b2454b3bceb45d7819cec9f000181574aabc621bb34d19893b5584"} Dec 03 06:45:28 crc kubenswrapper[4818]: E1203 06:45:28.059246 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e751691_c660_4527_a9a6_61505f83c6ff.slice/crio-303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0.scope\": RecentStats: unable to find data in memory cache]" Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.162759 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.286782 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d74d063c-c0e6-401c-857d-e5529f2df21a-scripts\") pod \"d74d063c-c0e6-401c-857d-e5529f2df21a\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.286952 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d74d063c-c0e6-401c-857d-e5529f2df21a-additional-scripts\") pod \"d74d063c-c0e6-401c-857d-e5529f2df21a\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.287004 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxhdq\" (UniqueName: \"kubernetes.io/projected/d74d063c-c0e6-401c-857d-e5529f2df21a-kube-api-access-jxhdq\") pod \"d74d063c-c0e6-401c-857d-e5529f2df21a\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.287041 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-run\") pod \"d74d063c-c0e6-401c-857d-e5529f2df21a\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.287070 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-log-ovn\") pod \"d74d063c-c0e6-401c-857d-e5529f2df21a\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.287126 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-run-ovn\") pod \"d74d063c-c0e6-401c-857d-e5529f2df21a\" (UID: \"d74d063c-c0e6-401c-857d-e5529f2df21a\") " Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.287213 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "d74d063c-c0e6-401c-857d-e5529f2df21a" (UID: "d74d063c-c0e6-401c-857d-e5529f2df21a"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.287380 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "d74d063c-c0e6-401c-857d-e5529f2df21a" (UID: "d74d063c-c0e6-401c-857d-e5529f2df21a"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.287979 4818 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.288012 4818 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.288029 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d74d063c-c0e6-401c-857d-e5529f2df21a-scripts" (OuterVolumeSpecName: "scripts") pod "d74d063c-c0e6-401c-857d-e5529f2df21a" (UID: "d74d063c-c0e6-401c-857d-e5529f2df21a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.288035 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-run" (OuterVolumeSpecName: "var-run") pod "d74d063c-c0e6-401c-857d-e5529f2df21a" (UID: "d74d063c-c0e6-401c-857d-e5529f2df21a"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.291415 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d74d063c-c0e6-401c-857d-e5529f2df21a-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "d74d063c-c0e6-401c-857d-e5529f2df21a" (UID: "d74d063c-c0e6-401c-857d-e5529f2df21a"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.293979 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d74d063c-c0e6-401c-857d-e5529f2df21a-kube-api-access-jxhdq" (OuterVolumeSpecName: "kube-api-access-jxhdq") pod "d74d063c-c0e6-401c-857d-e5529f2df21a" (UID: "d74d063c-c0e6-401c-857d-e5529f2df21a"). InnerVolumeSpecName "kube-api-access-jxhdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.389705 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxhdq\" (UniqueName: \"kubernetes.io/projected/d74d063c-c0e6-401c-857d-e5529f2df21a-kube-api-access-jxhdq\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.389750 4818 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d74d063c-c0e6-401c-857d-e5529f2df21a-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.389763 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d74d063c-c0e6-401c-857d-e5529f2df21a-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.389775 4818 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d74d063c-c0e6-401c-857d-e5529f2df21a-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.699480 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kvg5t-config-qtwrm" Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.699500 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kvg5t-config-qtwrm" event={"ID":"d74d063c-c0e6-401c-857d-e5529f2df21a","Type":"ContainerDied","Data":"8697403d833dd9d3174cc79098e7da3ce015633580a91a756cbd9de80656eb59"} Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.699881 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8697403d833dd9d3174cc79098e7da3ce015633580a91a756cbd9de80656eb59" Dec 03 06:45:28 crc kubenswrapper[4818]: I1203 06:45:28.703200 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f","Type":"ContainerStarted","Data":"13939c4a3ca2b73a0a96c8601e3a806ed0286b6cc3ce9cb6d434cab74e7eba2c"} Dec 03 06:45:29 crc kubenswrapper[4818]: I1203 06:45:29.286523 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-kvg5t-config-qtwrm"] Dec 03 06:45:29 crc kubenswrapper[4818]: I1203 06:45:29.293389 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-kvg5t-config-qtwrm"] Dec 03 06:45:29 crc kubenswrapper[4818]: I1203 06:45:29.726739 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f","Type":"ContainerStarted","Data":"68514bf97bdbd3f374ad5119a029e9185f6882444a12f6fdc0783565b7724e20"} Dec 03 06:45:29 crc kubenswrapper[4818]: I1203 06:45:29.726786 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f","Type":"ContainerStarted","Data":"b57431d718c97abfc9087875b7fdff2f5f6f9e8ddc7c41f0de0de0b4d625499d"} Dec 03 06:45:29 crc kubenswrapper[4818]: I1203 06:45:29.726800 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f","Type":"ContainerStarted","Data":"d815af22c4bbd8e5ae96d3239c306988ce98376e5ec950fd9b56e004be5e7f43"} Dec 03 06:45:30 crc kubenswrapper[4818]: I1203 06:45:30.542057 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-kvg5t" Dec 03 06:45:30 crc kubenswrapper[4818]: I1203 06:45:30.747307 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d74d063c-c0e6-401c-857d-e5529f2df21a" path="/var/lib/kubelet/pods/d74d063c-c0e6-401c-857d-e5529f2df21a/volumes" Dec 03 06:45:30 crc kubenswrapper[4818]: I1203 06:45:30.747954 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f","Type":"ContainerStarted","Data":"a04b0e11c00f69be63856b08c082a27b4a19d639053412d9837dc47b0a192506"} Dec 03 06:45:31 crc kubenswrapper[4818]: I1203 06:45:31.779958 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f","Type":"ContainerStarted","Data":"80538fa901a5d208d19ea133b13f891a6de1a23627a21d8299f262092767dcbd"} Dec 03 06:45:31 crc kubenswrapper[4818]: I1203 06:45:31.780271 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f","Type":"ContainerStarted","Data":"5d0d09089e5ba550461b56325c6d2fa5d47aea796d854d8519bf3e6ab940f32e"} Dec 03 06:45:31 crc kubenswrapper[4818]: I1203 06:45:31.780281 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f","Type":"ContainerStarted","Data":"f86eaaef508b29eb86fb877b8aad6afe81c9e5de67ca315044615432c38a6607"} Dec 03 06:45:32 crc kubenswrapper[4818]: I1203 06:45:32.792642 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f","Type":"ContainerStarted","Data":"bd0f950b58bc8a8ad43a641b795f6cf80a2f46632ac6d7eafd38ad9ffaacb7db"} Dec 03 06:45:32 crc kubenswrapper[4818]: I1203 06:45:32.793112 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f","Type":"ContainerStarted","Data":"6af518aa93d7d222bcc330333e147c1a7c8a2ec71e91a77cd44a198f2067ad79"} Dec 03 06:45:32 crc kubenswrapper[4818]: I1203 06:45:32.793132 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f","Type":"ContainerStarted","Data":"13cd45d17ba1c2a7dc5ff0c44fe2fbbb81db9105aa099bce7ae65f91bedba370"} Dec 03 06:45:32 crc kubenswrapper[4818]: I1203 06:45:32.793147 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"520a6b29-d2d4-46e4-a5ed-989f2fa1df1f","Type":"ContainerStarted","Data":"de882964627879550de4502600708d59825544e2654cd9c3fffea4f0b4514274"} Dec 03 06:45:32 crc kubenswrapper[4818]: I1203 06:45:32.832819 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.891268772 podStartE2EDuration="41.832794553s" podCreationTimestamp="2025-12-03 06:44:51 +0000 UTC" firstStartedPulling="2025-12-03 06:45:25.968783363 +0000 UTC m=+1083.660392135" lastFinishedPulling="2025-12-03 06:45:30.910309144 +0000 UTC m=+1088.601917916" observedRunningTime="2025-12-03 06:45:32.824461337 +0000 UTC m=+1090.516070099" watchObservedRunningTime="2025-12-03 06:45:32.832794553 +0000 UTC m=+1090.524403305" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.114656 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-8vdnp"] Dec 03 06:45:33 crc kubenswrapper[4818]: E1203 06:45:33.115063 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d74d063c-c0e6-401c-857d-e5529f2df21a" containerName="ovn-config" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.115088 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="d74d063c-c0e6-401c-857d-e5529f2df21a" containerName="ovn-config" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.115309 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="d74d063c-c0e6-401c-857d-e5529f2df21a" containerName="ovn-config" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.116340 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.118043 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.133838 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-8vdnp"] Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.236069 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.236478 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xml2m\" (UniqueName: \"kubernetes.io/projected/b9e387e8-8f19-4d97-834d-639710ce55dc-kube-api-access-xml2m\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.236519 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-dns-svc\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.236624 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-config\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.236687 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.236724 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.338077 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-dns-svc\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.338172 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-config\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.338231 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.338266 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.338313 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.338375 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xml2m\" (UniqueName: \"kubernetes.io/projected/b9e387e8-8f19-4d97-834d-639710ce55dc-kube-api-access-xml2m\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.339291 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.339656 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.339863 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.339948 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-config\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.340483 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-dns-svc\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.370708 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xml2m\" (UniqueName: \"kubernetes.io/projected/b9e387e8-8f19-4d97-834d-639710ce55dc-kube-api-access-xml2m\") pod \"dnsmasq-dns-764c5664d7-8vdnp\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.433429 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:33 crc kubenswrapper[4818]: I1203 06:45:33.882808 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-8vdnp"] Dec 03 06:45:34 crc kubenswrapper[4818]: I1203 06:45:34.811227 4818 generic.go:334] "Generic (PLEG): container finished" podID="b9e387e8-8f19-4d97-834d-639710ce55dc" containerID="e6ffe82240f0f0b63dd24e06c01f648b3492d61341cacdb4e187510f33f77bcf" exitCode=0 Dec 03 06:45:34 crc kubenswrapper[4818]: I1203 06:45:34.811328 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" event={"ID":"b9e387e8-8f19-4d97-834d-639710ce55dc","Type":"ContainerDied","Data":"e6ffe82240f0f0b63dd24e06c01f648b3492d61341cacdb4e187510f33f77bcf"} Dec 03 06:45:34 crc kubenswrapper[4818]: I1203 06:45:34.811597 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" event={"ID":"b9e387e8-8f19-4d97-834d-639710ce55dc","Type":"ContainerStarted","Data":"b6481b5ccf209098ca0ff0be372455fd1bcc7cd4060a6c33b396bcb80d18d02a"} Dec 03 06:45:35 crc kubenswrapper[4818]: I1203 06:45:35.882435 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 06:45:35 crc kubenswrapper[4818]: I1203 06:45:35.888845 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" event={"ID":"b9e387e8-8f19-4d97-834d-639710ce55dc","Type":"ContainerStarted","Data":"ffdc02851a91a239abe350b548ba5e893811da1284cbdc8d92949e14699a1b19"} Dec 03 06:45:35 crc kubenswrapper[4818]: I1203 06:45:35.888989 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:35 crc kubenswrapper[4818]: I1203 06:45:35.902134 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="0e751691-c660-4527-a9a6-61505f83c6ff" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Dec 03 06:45:35 crc kubenswrapper[4818]: I1203 06:45:35.902576 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="0e751691-c660-4527-a9a6-61505f83c6ff" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Dec 03 06:45:35 crc kubenswrapper[4818]: I1203 06:45:35.918970 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" podStartSLOduration=2.918946955 podStartE2EDuration="2.918946955s" podCreationTimestamp="2025-12-03 06:45:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:45:35.909750098 +0000 UTC m=+1093.601358870" watchObservedRunningTime="2025-12-03 06:45:35.918946955 +0000 UTC m=+1093.610555707" Dec 03 06:45:36 crc kubenswrapper[4818]: I1203 06:45:36.259031 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:45:38 crc kubenswrapper[4818]: E1203 06:45:38.291554 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e751691_c660_4527_a9a6_61505f83c6ff.slice/crio-303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0.scope\": RecentStats: unable to find data in memory cache]" Dec 03 06:45:39 crc kubenswrapper[4818]: I1203 06:45:39.927099 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-p57tz" event={"ID":"ba5ae97f-71e8-4dc1-bcd4-673e558c9693","Type":"ContainerStarted","Data":"cffcb4d6a6c778c5a1b0f933275c3fa74253fca3bf390eb0e427156485d8b910"} Dec 03 06:45:39 crc kubenswrapper[4818]: I1203 06:45:39.946841 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-p57tz" podStartSLOduration=2.408390202 podStartE2EDuration="30.946805794s" podCreationTimestamp="2025-12-03 06:45:09 +0000 UTC" firstStartedPulling="2025-12-03 06:45:10.603634662 +0000 UTC m=+1068.295243414" lastFinishedPulling="2025-12-03 06:45:39.142050244 +0000 UTC m=+1096.833659006" observedRunningTime="2025-12-03 06:45:39.941881862 +0000 UTC m=+1097.633490614" watchObservedRunningTime="2025-12-03 06:45:39.946805794 +0000 UTC m=+1097.638414546" Dec 03 06:45:43 crc kubenswrapper[4818]: I1203 06:45:43.434913 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:45:43 crc kubenswrapper[4818]: I1203 06:45:43.497154 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-8p9jz"] Dec 03 06:45:43 crc kubenswrapper[4818]: I1203 06:45:43.497433 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-8p9jz" podUID="6bf0813f-9f49-43b7-9e9a-8712e6a86159" containerName="dnsmasq-dns" containerID="cri-o://42b5eb85bf09ae4c51edd5147653b5843bc2cf0821155b35637d68d685905682" gracePeriod=10 Dec 03 06:45:43 crc kubenswrapper[4818]: I1203 06:45:43.911131 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:45:43 crc kubenswrapper[4818]: I1203 06:45:43.963090 4818 generic.go:334] "Generic (PLEG): container finished" podID="6bf0813f-9f49-43b7-9e9a-8712e6a86159" containerID="42b5eb85bf09ae4c51edd5147653b5843bc2cf0821155b35637d68d685905682" exitCode=0 Dec 03 06:45:43 crc kubenswrapper[4818]: I1203 06:45:43.963145 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-8p9jz" event={"ID":"6bf0813f-9f49-43b7-9e9a-8712e6a86159","Type":"ContainerDied","Data":"42b5eb85bf09ae4c51edd5147653b5843bc2cf0821155b35637d68d685905682"} Dec 03 06:45:43 crc kubenswrapper[4818]: I1203 06:45:43.963154 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-8p9jz" Dec 03 06:45:43 crc kubenswrapper[4818]: I1203 06:45:43.963170 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-8p9jz" event={"ID":"6bf0813f-9f49-43b7-9e9a-8712e6a86159","Type":"ContainerDied","Data":"981950291dede3a915ad46ea7bce9ce057bb3e581fd0412c1149d4e81019c50d"} Dec 03 06:45:43 crc kubenswrapper[4818]: I1203 06:45:43.963185 4818 scope.go:117] "RemoveContainer" containerID="42b5eb85bf09ae4c51edd5147653b5843bc2cf0821155b35637d68d685905682" Dec 03 06:45:43 crc kubenswrapper[4818]: I1203 06:45:43.980780 4818 scope.go:117] "RemoveContainer" containerID="0e35790c4bdc752e3a756f6b49e3294817e3332562c1a893fdf5ca8bf928fbed" Dec 03 06:45:43 crc kubenswrapper[4818]: I1203 06:45:43.997467 4818 scope.go:117] "RemoveContainer" containerID="42b5eb85bf09ae4c51edd5147653b5843bc2cf0821155b35637d68d685905682" Dec 03 06:45:44 crc kubenswrapper[4818]: E1203 06:45:44.003289 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42b5eb85bf09ae4c51edd5147653b5843bc2cf0821155b35637d68d685905682\": container with ID starting with 42b5eb85bf09ae4c51edd5147653b5843bc2cf0821155b35637d68d685905682 not found: ID does not exist" containerID="42b5eb85bf09ae4c51edd5147653b5843bc2cf0821155b35637d68d685905682" Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.003364 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42b5eb85bf09ae4c51edd5147653b5843bc2cf0821155b35637d68d685905682"} err="failed to get container status \"42b5eb85bf09ae4c51edd5147653b5843bc2cf0821155b35637d68d685905682\": rpc error: code = NotFound desc = could not find container \"42b5eb85bf09ae4c51edd5147653b5843bc2cf0821155b35637d68d685905682\": container with ID starting with 42b5eb85bf09ae4c51edd5147653b5843bc2cf0821155b35637d68d685905682 not found: ID does not exist" Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.003391 4818 scope.go:117] "RemoveContainer" containerID="0e35790c4bdc752e3a756f6b49e3294817e3332562c1a893fdf5ca8bf928fbed" Dec 03 06:45:44 crc kubenswrapper[4818]: E1203 06:45:44.003644 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e35790c4bdc752e3a756f6b49e3294817e3332562c1a893fdf5ca8bf928fbed\": container with ID starting with 0e35790c4bdc752e3a756f6b49e3294817e3332562c1a893fdf5ca8bf928fbed not found: ID does not exist" containerID="0e35790c4bdc752e3a756f6b49e3294817e3332562c1a893fdf5ca8bf928fbed" Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.003660 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e35790c4bdc752e3a756f6b49e3294817e3332562c1a893fdf5ca8bf928fbed"} err="failed to get container status \"0e35790c4bdc752e3a756f6b49e3294817e3332562c1a893fdf5ca8bf928fbed\": rpc error: code = NotFound desc = could not find container \"0e35790c4bdc752e3a756f6b49e3294817e3332562c1a893fdf5ca8bf928fbed\": container with ID starting with 0e35790c4bdc752e3a756f6b49e3294817e3332562c1a893fdf5ca8bf928fbed not found: ID does not exist" Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.079087 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82hjq\" (UniqueName: \"kubernetes.io/projected/6bf0813f-9f49-43b7-9e9a-8712e6a86159-kube-api-access-82hjq\") pod \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.079174 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-dns-svc\") pod \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.079205 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-ovsdbserver-nb\") pod \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.079248 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-config\") pod \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.079384 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-ovsdbserver-sb\") pod \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\" (UID: \"6bf0813f-9f49-43b7-9e9a-8712e6a86159\") " Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.085558 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bf0813f-9f49-43b7-9e9a-8712e6a86159-kube-api-access-82hjq" (OuterVolumeSpecName: "kube-api-access-82hjq") pod "6bf0813f-9f49-43b7-9e9a-8712e6a86159" (UID: "6bf0813f-9f49-43b7-9e9a-8712e6a86159"). InnerVolumeSpecName "kube-api-access-82hjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.123798 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-config" (OuterVolumeSpecName: "config") pod "6bf0813f-9f49-43b7-9e9a-8712e6a86159" (UID: "6bf0813f-9f49-43b7-9e9a-8712e6a86159"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.126307 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6bf0813f-9f49-43b7-9e9a-8712e6a86159" (UID: "6bf0813f-9f49-43b7-9e9a-8712e6a86159"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.126583 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6bf0813f-9f49-43b7-9e9a-8712e6a86159" (UID: "6bf0813f-9f49-43b7-9e9a-8712e6a86159"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.130455 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6bf0813f-9f49-43b7-9e9a-8712e6a86159" (UID: "6bf0813f-9f49-43b7-9e9a-8712e6a86159"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.181050 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.181094 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.181113 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82hjq\" (UniqueName: \"kubernetes.io/projected/6bf0813f-9f49-43b7-9e9a-8712e6a86159-kube-api-access-82hjq\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.181126 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.181136 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6bf0813f-9f49-43b7-9e9a-8712e6a86159-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.306843 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-8p9jz"] Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.314628 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-8p9jz"] Dec 03 06:45:44 crc kubenswrapper[4818]: I1203 06:45:44.753696 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bf0813f-9f49-43b7-9e9a-8712e6a86159" path="/var/lib/kubelet/pods/6bf0813f-9f49-43b7-9e9a-8712e6a86159/volumes" Dec 03 06:45:45 crc kubenswrapper[4818]: I1203 06:45:45.868947 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 06:45:45 crc kubenswrapper[4818]: I1203 06:45:45.987738 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-p57tz" event={"ID":"ba5ae97f-71e8-4dc1-bcd4-673e558c9693","Type":"ContainerDied","Data":"cffcb4d6a6c778c5a1b0f933275c3fa74253fca3bf390eb0e427156485d8b910"} Dec 03 06:45:45 crc kubenswrapper[4818]: I1203 06:45:45.987745 4818 generic.go:334] "Generic (PLEG): container finished" podID="ba5ae97f-71e8-4dc1-bcd4-673e558c9693" containerID="cffcb4d6a6c778c5a1b0f933275c3fa74253fca3bf390eb0e427156485d8b910" exitCode=0 Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.219015 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-wpt88"] Dec 03 06:45:46 crc kubenswrapper[4818]: E1203 06:45:46.219379 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bf0813f-9f49-43b7-9e9a-8712e6a86159" containerName="dnsmasq-dns" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.219395 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bf0813f-9f49-43b7-9e9a-8712e6a86159" containerName="dnsmasq-dns" Dec 03 06:45:46 crc kubenswrapper[4818]: E1203 06:45:46.219432 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bf0813f-9f49-43b7-9e9a-8712e6a86159" containerName="init" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.219438 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bf0813f-9f49-43b7-9e9a-8712e6a86159" containerName="init" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.219586 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bf0813f-9f49-43b7-9e9a-8712e6a86159" containerName="dnsmasq-dns" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.220079 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wpt88" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.241627 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wpt88"] Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.249544 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-d39d-account-create-update-hjjxx"] Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.250988 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d39d-account-create-update-hjjxx" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.252947 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.263024 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d39d-account-create-update-hjjxx"] Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.316920 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8k2l\" (UniqueName: \"kubernetes.io/projected/388347b6-ce3b-4469-ad87-32faa60f8d21-kube-api-access-s8k2l\") pod \"cinder-db-create-wpt88\" (UID: \"388347b6-ce3b-4469-ad87-32faa60f8d21\") " pod="openstack/cinder-db-create-wpt88" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.316983 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0c0052f-9ec7-43ad-87a0-87a32f8461b7-operator-scripts\") pod \"cinder-d39d-account-create-update-hjjxx\" (UID: \"e0c0052f-9ec7-43ad-87a0-87a32f8461b7\") " pod="openstack/cinder-d39d-account-create-update-hjjxx" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.317066 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/388347b6-ce3b-4469-ad87-32faa60f8d21-operator-scripts\") pod \"cinder-db-create-wpt88\" (UID: \"388347b6-ce3b-4469-ad87-32faa60f8d21\") " pod="openstack/cinder-db-create-wpt88" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.317146 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkcbs\" (UniqueName: \"kubernetes.io/projected/e0c0052f-9ec7-43ad-87a0-87a32f8461b7-kube-api-access-qkcbs\") pod \"cinder-d39d-account-create-update-hjjxx\" (UID: \"e0c0052f-9ec7-43ad-87a0-87a32f8461b7\") " pod="openstack/cinder-d39d-account-create-update-hjjxx" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.319633 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-d7v9h"] Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.320697 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d7v9h" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.352785 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-d7v9h"] Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.418211 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8k2l\" (UniqueName: \"kubernetes.io/projected/388347b6-ce3b-4469-ad87-32faa60f8d21-kube-api-access-s8k2l\") pod \"cinder-db-create-wpt88\" (UID: \"388347b6-ce3b-4469-ad87-32faa60f8d21\") " pod="openstack/cinder-db-create-wpt88" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.418271 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0c0052f-9ec7-43ad-87a0-87a32f8461b7-operator-scripts\") pod \"cinder-d39d-account-create-update-hjjxx\" (UID: \"e0c0052f-9ec7-43ad-87a0-87a32f8461b7\") " pod="openstack/cinder-d39d-account-create-update-hjjxx" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.418340 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb6x6\" (UniqueName: \"kubernetes.io/projected/939f2f3e-bc03-445c-bc1e-2cb1ca895795-kube-api-access-fb6x6\") pod \"barbican-db-create-d7v9h\" (UID: \"939f2f3e-bc03-445c-bc1e-2cb1ca895795\") " pod="openstack/barbican-db-create-d7v9h" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.418391 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/388347b6-ce3b-4469-ad87-32faa60f8d21-operator-scripts\") pod \"cinder-db-create-wpt88\" (UID: \"388347b6-ce3b-4469-ad87-32faa60f8d21\") " pod="openstack/cinder-db-create-wpt88" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.418418 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/939f2f3e-bc03-445c-bc1e-2cb1ca895795-operator-scripts\") pod \"barbican-db-create-d7v9h\" (UID: \"939f2f3e-bc03-445c-bc1e-2cb1ca895795\") " pod="openstack/barbican-db-create-d7v9h" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.418466 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkcbs\" (UniqueName: \"kubernetes.io/projected/e0c0052f-9ec7-43ad-87a0-87a32f8461b7-kube-api-access-qkcbs\") pod \"cinder-d39d-account-create-update-hjjxx\" (UID: \"e0c0052f-9ec7-43ad-87a0-87a32f8461b7\") " pod="openstack/cinder-d39d-account-create-update-hjjxx" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.419767 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0c0052f-9ec7-43ad-87a0-87a32f8461b7-operator-scripts\") pod \"cinder-d39d-account-create-update-hjjxx\" (UID: \"e0c0052f-9ec7-43ad-87a0-87a32f8461b7\") " pod="openstack/cinder-d39d-account-create-update-hjjxx" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.419839 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/388347b6-ce3b-4469-ad87-32faa60f8d21-operator-scripts\") pod \"cinder-db-create-wpt88\" (UID: \"388347b6-ce3b-4469-ad87-32faa60f8d21\") " pod="openstack/cinder-db-create-wpt88" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.423604 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-0a8a-account-create-update-82q4f"] Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.424606 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0a8a-account-create-update-82q4f" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.434162 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.448883 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0a8a-account-create-update-82q4f"] Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.472804 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkcbs\" (UniqueName: \"kubernetes.io/projected/e0c0052f-9ec7-43ad-87a0-87a32f8461b7-kube-api-access-qkcbs\") pod \"cinder-d39d-account-create-update-hjjxx\" (UID: \"e0c0052f-9ec7-43ad-87a0-87a32f8461b7\") " pod="openstack/cinder-d39d-account-create-update-hjjxx" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.478964 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8k2l\" (UniqueName: \"kubernetes.io/projected/388347b6-ce3b-4469-ad87-32faa60f8d21-kube-api-access-s8k2l\") pod \"cinder-db-create-wpt88\" (UID: \"388347b6-ce3b-4469-ad87-32faa60f8d21\") " pod="openstack/cinder-db-create-wpt88" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.479628 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-p6k8k"] Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.480729 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-p6k8k" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.482600 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.482899 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.485132 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.485165 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ggd6v" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.492074 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-p6k8k"] Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.519799 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d32631e-bf3a-46d9-bfd3-2ea9e21987c0-operator-scripts\") pod \"barbican-0a8a-account-create-update-82q4f\" (UID: \"4d32631e-bf3a-46d9-bfd3-2ea9e21987c0\") " pod="openstack/barbican-0a8a-account-create-update-82q4f" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.519881 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb6x6\" (UniqueName: \"kubernetes.io/projected/939f2f3e-bc03-445c-bc1e-2cb1ca895795-kube-api-access-fb6x6\") pod \"barbican-db-create-d7v9h\" (UID: \"939f2f3e-bc03-445c-bc1e-2cb1ca895795\") " pod="openstack/barbican-db-create-d7v9h" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.519938 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/939f2f3e-bc03-445c-bc1e-2cb1ca895795-operator-scripts\") pod \"barbican-db-create-d7v9h\" (UID: \"939f2f3e-bc03-445c-bc1e-2cb1ca895795\") " pod="openstack/barbican-db-create-d7v9h" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.519995 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h78lq\" (UniqueName: \"kubernetes.io/projected/4d32631e-bf3a-46d9-bfd3-2ea9e21987c0-kube-api-access-h78lq\") pod \"barbican-0a8a-account-create-update-82q4f\" (UID: \"4d32631e-bf3a-46d9-bfd3-2ea9e21987c0\") " pod="openstack/barbican-0a8a-account-create-update-82q4f" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.520806 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/939f2f3e-bc03-445c-bc1e-2cb1ca895795-operator-scripts\") pod \"barbican-db-create-d7v9h\" (UID: \"939f2f3e-bc03-445c-bc1e-2cb1ca895795\") " pod="openstack/barbican-db-create-d7v9h" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.534880 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb6x6\" (UniqueName: \"kubernetes.io/projected/939f2f3e-bc03-445c-bc1e-2cb1ca895795-kube-api-access-fb6x6\") pod \"barbican-db-create-d7v9h\" (UID: \"939f2f3e-bc03-445c-bc1e-2cb1ca895795\") " pod="openstack/barbican-db-create-d7v9h" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.537021 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wpt88" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.571142 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d39d-account-create-update-hjjxx" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.621319 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed38224f-53f7-4e63-8841-4d1c3f65c456-config-data\") pod \"keystone-db-sync-p6k8k\" (UID: \"ed38224f-53f7-4e63-8841-4d1c3f65c456\") " pod="openstack/keystone-db-sync-p6k8k" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.621669 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d32631e-bf3a-46d9-bfd3-2ea9e21987c0-operator-scripts\") pod \"barbican-0a8a-account-create-update-82q4f\" (UID: \"4d32631e-bf3a-46d9-bfd3-2ea9e21987c0\") " pod="openstack/barbican-0a8a-account-create-update-82q4f" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.621711 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf55t\" (UniqueName: \"kubernetes.io/projected/ed38224f-53f7-4e63-8841-4d1c3f65c456-kube-api-access-hf55t\") pod \"keystone-db-sync-p6k8k\" (UID: \"ed38224f-53f7-4e63-8841-4d1c3f65c456\") " pod="openstack/keystone-db-sync-p6k8k" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.621734 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed38224f-53f7-4e63-8841-4d1c3f65c456-combined-ca-bundle\") pod \"keystone-db-sync-p6k8k\" (UID: \"ed38224f-53f7-4e63-8841-4d1c3f65c456\") " pod="openstack/keystone-db-sync-p6k8k" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.621776 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h78lq\" (UniqueName: \"kubernetes.io/projected/4d32631e-bf3a-46d9-bfd3-2ea9e21987c0-kube-api-access-h78lq\") pod \"barbican-0a8a-account-create-update-82q4f\" (UID: \"4d32631e-bf3a-46d9-bfd3-2ea9e21987c0\") " pod="openstack/barbican-0a8a-account-create-update-82q4f" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.622741 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d32631e-bf3a-46d9-bfd3-2ea9e21987c0-operator-scripts\") pod \"barbican-0a8a-account-create-update-82q4f\" (UID: \"4d32631e-bf3a-46d9-bfd3-2ea9e21987c0\") " pod="openstack/barbican-0a8a-account-create-update-82q4f" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.633521 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-pdz8d"] Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.634616 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pdz8d" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.641896 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d7v9h" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.651877 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h78lq\" (UniqueName: \"kubernetes.io/projected/4d32631e-bf3a-46d9-bfd3-2ea9e21987c0-kube-api-access-h78lq\") pod \"barbican-0a8a-account-create-update-82q4f\" (UID: \"4d32631e-bf3a-46d9-bfd3-2ea9e21987c0\") " pod="openstack/barbican-0a8a-account-create-update-82q4f" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.745191 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-339d-account-create-update-5hpch"] Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.749109 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0a8a-account-create-update-82q4f" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.766117 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-339d-account-create-update-5hpch" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.788257 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.788278 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed38224f-53f7-4e63-8841-4d1c3f65c456-config-data\") pod \"keystone-db-sync-p6k8k\" (UID: \"ed38224f-53f7-4e63-8841-4d1c3f65c456\") " pod="openstack/keystone-db-sync-p6k8k" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.788863 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf55t\" (UniqueName: \"kubernetes.io/projected/ed38224f-53f7-4e63-8841-4d1c3f65c456-kube-api-access-hf55t\") pod \"keystone-db-sync-p6k8k\" (UID: \"ed38224f-53f7-4e63-8841-4d1c3f65c456\") " pod="openstack/keystone-db-sync-p6k8k" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.788965 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed38224f-53f7-4e63-8841-4d1c3f65c456-combined-ca-bundle\") pod \"keystone-db-sync-p6k8k\" (UID: \"ed38224f-53f7-4e63-8841-4d1c3f65c456\") " pod="openstack/keystone-db-sync-p6k8k" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.789066 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/038a0b6a-32a3-407f-9e31-0f133592cc57-operator-scripts\") pod \"neutron-db-create-pdz8d\" (UID: \"038a0b6a-32a3-407f-9e31-0f133592cc57\") " pod="openstack/neutron-db-create-pdz8d" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.789225 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlx9w\" (UniqueName: \"kubernetes.io/projected/038a0b6a-32a3-407f-9e31-0f133592cc57-kube-api-access-dlx9w\") pod \"neutron-db-create-pdz8d\" (UID: \"038a0b6a-32a3-407f-9e31-0f133592cc57\") " pod="openstack/neutron-db-create-pdz8d" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.793318 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed38224f-53f7-4e63-8841-4d1c3f65c456-config-data\") pod \"keystone-db-sync-p6k8k\" (UID: \"ed38224f-53f7-4e63-8841-4d1c3f65c456\") " pod="openstack/keystone-db-sync-p6k8k" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.809564 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed38224f-53f7-4e63-8841-4d1c3f65c456-combined-ca-bundle\") pod \"keystone-db-sync-p6k8k\" (UID: \"ed38224f-53f7-4e63-8841-4d1c3f65c456\") " pod="openstack/keystone-db-sync-p6k8k" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.816296 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf55t\" (UniqueName: \"kubernetes.io/projected/ed38224f-53f7-4e63-8841-4d1c3f65c456-kube-api-access-hf55t\") pod \"keystone-db-sync-p6k8k\" (UID: \"ed38224f-53f7-4e63-8841-4d1c3f65c456\") " pod="openstack/keystone-db-sync-p6k8k" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.823364 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-339d-account-create-update-5hpch"] Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.823573 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-pdz8d"] Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.891004 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8063467-b639-4aae-8d0d-db6660dfadac-operator-scripts\") pod \"neutron-339d-account-create-update-5hpch\" (UID: \"c8063467-b639-4aae-8d0d-db6660dfadac\") " pod="openstack/neutron-339d-account-create-update-5hpch" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.891107 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/038a0b6a-32a3-407f-9e31-0f133592cc57-operator-scripts\") pod \"neutron-db-create-pdz8d\" (UID: \"038a0b6a-32a3-407f-9e31-0f133592cc57\") " pod="openstack/neutron-db-create-pdz8d" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.891151 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqk5c\" (UniqueName: \"kubernetes.io/projected/c8063467-b639-4aae-8d0d-db6660dfadac-kube-api-access-sqk5c\") pod \"neutron-339d-account-create-update-5hpch\" (UID: \"c8063467-b639-4aae-8d0d-db6660dfadac\") " pod="openstack/neutron-339d-account-create-update-5hpch" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.891167 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlx9w\" (UniqueName: \"kubernetes.io/projected/038a0b6a-32a3-407f-9e31-0f133592cc57-kube-api-access-dlx9w\") pod \"neutron-db-create-pdz8d\" (UID: \"038a0b6a-32a3-407f-9e31-0f133592cc57\") " pod="openstack/neutron-db-create-pdz8d" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.892281 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/038a0b6a-32a3-407f-9e31-0f133592cc57-operator-scripts\") pod \"neutron-db-create-pdz8d\" (UID: \"038a0b6a-32a3-407f-9e31-0f133592cc57\") " pod="openstack/neutron-db-create-pdz8d" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.920440 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlx9w\" (UniqueName: \"kubernetes.io/projected/038a0b6a-32a3-407f-9e31-0f133592cc57-kube-api-access-dlx9w\") pod \"neutron-db-create-pdz8d\" (UID: \"038a0b6a-32a3-407f-9e31-0f133592cc57\") " pod="openstack/neutron-db-create-pdz8d" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.938311 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-p6k8k" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.949375 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pdz8d" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.992228 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqk5c\" (UniqueName: \"kubernetes.io/projected/c8063467-b639-4aae-8d0d-db6660dfadac-kube-api-access-sqk5c\") pod \"neutron-339d-account-create-update-5hpch\" (UID: \"c8063467-b639-4aae-8d0d-db6660dfadac\") " pod="openstack/neutron-339d-account-create-update-5hpch" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.992416 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8063467-b639-4aae-8d0d-db6660dfadac-operator-scripts\") pod \"neutron-339d-account-create-update-5hpch\" (UID: \"c8063467-b639-4aae-8d0d-db6660dfadac\") " pod="openstack/neutron-339d-account-create-update-5hpch" Dec 03 06:45:46 crc kubenswrapper[4818]: I1203 06:45:46.993405 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8063467-b639-4aae-8d0d-db6660dfadac-operator-scripts\") pod \"neutron-339d-account-create-update-5hpch\" (UID: \"c8063467-b639-4aae-8d0d-db6660dfadac\") " pod="openstack/neutron-339d-account-create-update-5hpch" Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.018443 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqk5c\" (UniqueName: \"kubernetes.io/projected/c8063467-b639-4aae-8d0d-db6660dfadac-kube-api-access-sqk5c\") pod \"neutron-339d-account-create-update-5hpch\" (UID: \"c8063467-b639-4aae-8d0d-db6660dfadac\") " pod="openstack/neutron-339d-account-create-update-5hpch" Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.058611 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wpt88"] Dec 03 06:45:47 crc kubenswrapper[4818]: W1203 06:45:47.075213 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod388347b6_ce3b_4469_ad87_32faa60f8d21.slice/crio-78715f476929389945cb38a8a8c6e374b065604d8b4e8b8d8d323f731944baa7 WatchSource:0}: Error finding container 78715f476929389945cb38a8a8c6e374b065604d8b4e8b8d8d323f731944baa7: Status 404 returned error can't find the container with id 78715f476929389945cb38a8a8c6e374b065604d8b4e8b8d8d323f731944baa7 Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.106407 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d39d-account-create-update-hjjxx"] Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.125230 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-339d-account-create-update-5hpch" Dec 03 06:45:47 crc kubenswrapper[4818]: W1203 06:45:47.125654 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0c0052f_9ec7_43ad_87a0_87a32f8461b7.slice/crio-de34c4e970cb1f5de61c50ea1199585e408dc2585cdeb4f4069e394b2946f8da WatchSource:0}: Error finding container de34c4e970cb1f5de61c50ea1199585e408dc2585cdeb4f4069e394b2946f8da: Status 404 returned error can't find the container with id de34c4e970cb1f5de61c50ea1199585e408dc2585cdeb4f4069e394b2946f8da Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.394578 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0a8a-account-create-update-82q4f"] Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.547769 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-d7v9h"] Dec 03 06:45:47 crc kubenswrapper[4818]: W1203 06:45:47.560450 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod939f2f3e_bc03_445c_bc1e_2cb1ca895795.slice/crio-907ebae80518454f8b34ac0aad092dcb15398d322cc7159da4c3c879a0a7b2aa WatchSource:0}: Error finding container 907ebae80518454f8b34ac0aad092dcb15398d322cc7159da4c3c879a0a7b2aa: Status 404 returned error can't find the container with id 907ebae80518454f8b34ac0aad092dcb15398d322cc7159da4c3c879a0a7b2aa Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.703747 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-pdz8d"] Dec 03 06:45:47 crc kubenswrapper[4818]: W1203 06:45:47.707974 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod038a0b6a_32a3_407f_9e31_0f133592cc57.slice/crio-321a89d76c01759fd51f8175ee88510ce86ce103e037f31ce238a5424439df44 WatchSource:0}: Error finding container 321a89d76c01759fd51f8175ee88510ce86ce103e037f31ce238a5424439df44: Status 404 returned error can't find the container with id 321a89d76c01759fd51f8175ee88510ce86ce103e037f31ce238a5424439df44 Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.749915 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-p57tz" Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.846542 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-p6k8k"] Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.900011 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-339d-account-create-update-5hpch"] Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.915789 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-combined-ca-bundle\") pod \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\" (UID: \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\") " Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.915914 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-config-data\") pod \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\" (UID: \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\") " Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.915955 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zld56\" (UniqueName: \"kubernetes.io/projected/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-kube-api-access-zld56\") pod \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\" (UID: \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\") " Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.915979 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-db-sync-config-data\") pod \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\" (UID: \"ba5ae97f-71e8-4dc1-bcd4-673e558c9693\") " Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.924963 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ba5ae97f-71e8-4dc1-bcd4-673e558c9693" (UID: "ba5ae97f-71e8-4dc1-bcd4-673e558c9693"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.925122 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-kube-api-access-zld56" (OuterVolumeSpecName: "kube-api-access-zld56") pod "ba5ae97f-71e8-4dc1-bcd4-673e558c9693" (UID: "ba5ae97f-71e8-4dc1-bcd4-673e558c9693"). InnerVolumeSpecName "kube-api-access-zld56". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:47 crc kubenswrapper[4818]: W1203 06:45:47.939501 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8063467_b639_4aae_8d0d_db6660dfadac.slice/crio-c581b828a2afd62a83718b90915aac028803273e9ca2efccf0639aec4218c882 WatchSource:0}: Error finding container c581b828a2afd62a83718b90915aac028803273e9ca2efccf0639aec4218c882: Status 404 returned error can't find the container with id c581b828a2afd62a83718b90915aac028803273e9ca2efccf0639aec4218c882 Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.945046 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ba5ae97f-71e8-4dc1-bcd4-673e558c9693" (UID: "ba5ae97f-71e8-4dc1-bcd4-673e558c9693"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:45:47 crc kubenswrapper[4818]: I1203 06:45:47.994986 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-config-data" (OuterVolumeSpecName: "config-data") pod "ba5ae97f-71e8-4dc1-bcd4-673e558c9693" (UID: "ba5ae97f-71e8-4dc1-bcd4-673e558c9693"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.174218 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.174278 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.174302 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zld56\" (UniqueName: \"kubernetes.io/projected/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-kube-api-access-zld56\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.174315 4818 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ba5ae97f-71e8-4dc1-bcd4-673e558c9693-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.195402 4818 generic.go:334] "Generic (PLEG): container finished" podID="e0c0052f-9ec7-43ad-87a0-87a32f8461b7" containerID="29688ec029f0eab34d50b274713df996431a8784d7b5298c51eeb3d967467d2c" exitCode=0 Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.195804 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d39d-account-create-update-hjjxx" event={"ID":"e0c0052f-9ec7-43ad-87a0-87a32f8461b7","Type":"ContainerDied","Data":"29688ec029f0eab34d50b274713df996431a8784d7b5298c51eeb3d967467d2c"} Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.195848 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d39d-account-create-update-hjjxx" event={"ID":"e0c0052f-9ec7-43ad-87a0-87a32f8461b7","Type":"ContainerStarted","Data":"de34c4e970cb1f5de61c50ea1199585e408dc2585cdeb4f4069e394b2946f8da"} Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.204202 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pdz8d" event={"ID":"038a0b6a-32a3-407f-9e31-0f133592cc57","Type":"ContainerStarted","Data":"8f6fae7c4ae109b8c787819a2295aefc2e9b087d3e810b8bbcae9c2e838d0762"} Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.204234 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pdz8d" event={"ID":"038a0b6a-32a3-407f-9e31-0f133592cc57","Type":"ContainerStarted","Data":"321a89d76c01759fd51f8175ee88510ce86ce103e037f31ce238a5424439df44"} Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.207735 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-339d-account-create-update-5hpch" event={"ID":"c8063467-b639-4aae-8d0d-db6660dfadac","Type":"ContainerStarted","Data":"c581b828a2afd62a83718b90915aac028803273e9ca2efccf0639aec4218c882"} Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.209196 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-p6k8k" event={"ID":"ed38224f-53f7-4e63-8841-4d1c3f65c456","Type":"ContainerStarted","Data":"475139a7015ad51e6215c4bb1e1f0012056404367127ec7add4177f3e3f99d4b"} Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.214136 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d7v9h" event={"ID":"939f2f3e-bc03-445c-bc1e-2cb1ca895795","Type":"ContainerStarted","Data":"e915d4dd338656293c4c2bfdb99ff185eee65a4f09098f99f667aac19087fa8f"} Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.214181 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d7v9h" event={"ID":"939f2f3e-bc03-445c-bc1e-2cb1ca895795","Type":"ContainerStarted","Data":"907ebae80518454f8b34ac0aad092dcb15398d322cc7159da4c3c879a0a7b2aa"} Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.219858 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-p57tz" event={"ID":"ba5ae97f-71e8-4dc1-bcd4-673e558c9693","Type":"ContainerDied","Data":"a36f2ebc02e5ea46d3914af488b14b60e216658a93469404cb74da964c2379c9"} Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.220056 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a36f2ebc02e5ea46d3914af488b14b60e216658a93469404cb74da964c2379c9" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.219885 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-p57tz" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.221355 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0a8a-account-create-update-82q4f" event={"ID":"4d32631e-bf3a-46d9-bfd3-2ea9e21987c0","Type":"ContainerStarted","Data":"2a4cea9459d21bb4c38e442530d2daf52454d4c0750a2ae66afe12c7aa793625"} Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.221394 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0a8a-account-create-update-82q4f" event={"ID":"4d32631e-bf3a-46d9-bfd3-2ea9e21987c0","Type":"ContainerStarted","Data":"c4e44949771f1e325e0aa0932eb7f0afe87678b5d4d0579027ecd4846ef02617"} Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.228102 4818 generic.go:334] "Generic (PLEG): container finished" podID="388347b6-ce3b-4469-ad87-32faa60f8d21" containerID="030d8ac7ea51b4d702857b4ceb17fad2a36177d18d16e2b97820e6a3c67db8c0" exitCode=0 Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.228171 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wpt88" event={"ID":"388347b6-ce3b-4469-ad87-32faa60f8d21","Type":"ContainerDied","Data":"030d8ac7ea51b4d702857b4ceb17fad2a36177d18d16e2b97820e6a3c67db8c0"} Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.228211 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wpt88" event={"ID":"388347b6-ce3b-4469-ad87-32faa60f8d21","Type":"ContainerStarted","Data":"78715f476929389945cb38a8a8c6e374b065604d8b4e8b8d8d323f731944baa7"} Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.232792 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-pdz8d" podStartSLOduration=2.232776955 podStartE2EDuration="2.232776955s" podCreationTimestamp="2025-12-03 06:45:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:45:48.225565556 +0000 UTC m=+1105.917174308" watchObservedRunningTime="2025-12-03 06:45:48.232776955 +0000 UTC m=+1105.924385707" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.252483 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-0a8a-account-create-update-82q4f" podStartSLOduration=2.252462541 podStartE2EDuration="2.252462541s" podCreationTimestamp="2025-12-03 06:45:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:45:48.2455094 +0000 UTC m=+1105.937118152" watchObservedRunningTime="2025-12-03 06:45:48.252462541 +0000 UTC m=+1105.944071293" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.273551 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-d7v9h" podStartSLOduration=2.273532462 podStartE2EDuration="2.273532462s" podCreationTimestamp="2025-12-03 06:45:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:45:48.265802781 +0000 UTC m=+1105.957411533" watchObservedRunningTime="2025-12-03 06:45:48.273532462 +0000 UTC m=+1105.965141224" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.710425 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sfd82"] Dec 03 06:45:48 crc kubenswrapper[4818]: E1203 06:45:48.710754 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5ae97f-71e8-4dc1-bcd4-673e558c9693" containerName="glance-db-sync" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.710768 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5ae97f-71e8-4dc1-bcd4-673e558c9693" containerName="glance-db-sync" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.710958 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba5ae97f-71e8-4dc1-bcd4-673e558c9693" containerName="glance-db-sync" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.711714 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:48 crc kubenswrapper[4818]: E1203 06:45:48.719072 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d32631e_bf3a_46d9_bfd3_2ea9e21987c0.slice/crio-conmon-2a4cea9459d21bb4c38e442530d2daf52454d4c0750a2ae66afe12c7aa793625.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d32631e_bf3a_46d9_bfd3_2ea9e21987c0.slice/crio-2a4cea9459d21bb4c38e442530d2daf52454d4c0750a2ae66afe12c7aa793625.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e751691_c660_4527_a9a6_61505f83c6ff.slice/crio-303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0.scope\": RecentStats: unable to find data in memory cache]" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.804272 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sfd82"] Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.902973 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.905201 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.906136 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.906244 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdkrp\" (UniqueName: \"kubernetes.io/projected/4803b078-4593-4b55-8a42-f4c3f4b0e21f-kube-api-access-gdkrp\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.906306 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-config\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:48 crc kubenswrapper[4818]: I1203 06:45:48.906342 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.007608 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.007680 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.007714 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.007735 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdkrp\" (UniqueName: \"kubernetes.io/projected/4803b078-4593-4b55-8a42-f4c3f4b0e21f-kube-api-access-gdkrp\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.007759 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-config\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.007778 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.008772 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.009393 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.010132 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.010614 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-config\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.010929 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.030637 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdkrp\" (UniqueName: \"kubernetes.io/projected/4803b078-4593-4b55-8a42-f4c3f4b0e21f-kube-api-access-gdkrp\") pod \"dnsmasq-dns-74f6bcbc87-sfd82\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.083285 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.305859 4818 generic.go:334] "Generic (PLEG): container finished" podID="038a0b6a-32a3-407f-9e31-0f133592cc57" containerID="8f6fae7c4ae109b8c787819a2295aefc2e9b087d3e810b8bbcae9c2e838d0762" exitCode=0 Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.305944 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pdz8d" event={"ID":"038a0b6a-32a3-407f-9e31-0f133592cc57","Type":"ContainerDied","Data":"8f6fae7c4ae109b8c787819a2295aefc2e9b087d3e810b8bbcae9c2e838d0762"} Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.308577 4818 generic.go:334] "Generic (PLEG): container finished" podID="c8063467-b639-4aae-8d0d-db6660dfadac" containerID="0c0f22c08cd47528c691be66785e371daca73f95c3122d51f2830156c4a7c5a1" exitCode=0 Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.308639 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-339d-account-create-update-5hpch" event={"ID":"c8063467-b639-4aae-8d0d-db6660dfadac","Type":"ContainerDied","Data":"0c0f22c08cd47528c691be66785e371daca73f95c3122d51f2830156c4a7c5a1"} Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.311537 4818 generic.go:334] "Generic (PLEG): container finished" podID="939f2f3e-bc03-445c-bc1e-2cb1ca895795" containerID="e915d4dd338656293c4c2bfdb99ff185eee65a4f09098f99f667aac19087fa8f" exitCode=0 Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.311596 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d7v9h" event={"ID":"939f2f3e-bc03-445c-bc1e-2cb1ca895795","Type":"ContainerDied","Data":"e915d4dd338656293c4c2bfdb99ff185eee65a4f09098f99f667aac19087fa8f"} Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.316602 4818 generic.go:334] "Generic (PLEG): container finished" podID="4d32631e-bf3a-46d9-bfd3-2ea9e21987c0" containerID="2a4cea9459d21bb4c38e442530d2daf52454d4c0750a2ae66afe12c7aa793625" exitCode=0 Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.316801 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0a8a-account-create-update-82q4f" event={"ID":"4d32631e-bf3a-46d9-bfd3-2ea9e21987c0","Type":"ContainerDied","Data":"2a4cea9459d21bb4c38e442530d2daf52454d4c0750a2ae66afe12c7aa793625"} Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.692907 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sfd82"] Dec 03 06:45:49 crc kubenswrapper[4818]: W1203 06:45:49.696727 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4803b078_4593_4b55_8a42_f4c3f4b0e21f.slice/crio-99a4375a31a3b06f220595540cb9121692a2490d26e646ddea9c1148f74e88f1 WatchSource:0}: Error finding container 99a4375a31a3b06f220595540cb9121692a2490d26e646ddea9c1148f74e88f1: Status 404 returned error can't find the container with id 99a4375a31a3b06f220595540cb9121692a2490d26e646ddea9c1148f74e88f1 Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.848568 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d39d-account-create-update-hjjxx" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.858934 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wpt88" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.905339 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8k2l\" (UniqueName: \"kubernetes.io/projected/388347b6-ce3b-4469-ad87-32faa60f8d21-kube-api-access-s8k2l\") pod \"388347b6-ce3b-4469-ad87-32faa60f8d21\" (UID: \"388347b6-ce3b-4469-ad87-32faa60f8d21\") " Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.905409 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0c0052f-9ec7-43ad-87a0-87a32f8461b7-operator-scripts\") pod \"e0c0052f-9ec7-43ad-87a0-87a32f8461b7\" (UID: \"e0c0052f-9ec7-43ad-87a0-87a32f8461b7\") " Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.905579 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/388347b6-ce3b-4469-ad87-32faa60f8d21-operator-scripts\") pod \"388347b6-ce3b-4469-ad87-32faa60f8d21\" (UID: \"388347b6-ce3b-4469-ad87-32faa60f8d21\") " Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.905620 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkcbs\" (UniqueName: \"kubernetes.io/projected/e0c0052f-9ec7-43ad-87a0-87a32f8461b7-kube-api-access-qkcbs\") pod \"e0c0052f-9ec7-43ad-87a0-87a32f8461b7\" (UID: \"e0c0052f-9ec7-43ad-87a0-87a32f8461b7\") " Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.906859 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/388347b6-ce3b-4469-ad87-32faa60f8d21-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "388347b6-ce3b-4469-ad87-32faa60f8d21" (UID: "388347b6-ce3b-4469-ad87-32faa60f8d21"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.907354 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0c0052f-9ec7-43ad-87a0-87a32f8461b7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e0c0052f-9ec7-43ad-87a0-87a32f8461b7" (UID: "e0c0052f-9ec7-43ad-87a0-87a32f8461b7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.910346 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0c0052f-9ec7-43ad-87a0-87a32f8461b7-kube-api-access-qkcbs" (OuterVolumeSpecName: "kube-api-access-qkcbs") pod "e0c0052f-9ec7-43ad-87a0-87a32f8461b7" (UID: "e0c0052f-9ec7-43ad-87a0-87a32f8461b7"). InnerVolumeSpecName "kube-api-access-qkcbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:49 crc kubenswrapper[4818]: I1203 06:45:49.911622 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/388347b6-ce3b-4469-ad87-32faa60f8d21-kube-api-access-s8k2l" (OuterVolumeSpecName: "kube-api-access-s8k2l") pod "388347b6-ce3b-4469-ad87-32faa60f8d21" (UID: "388347b6-ce3b-4469-ad87-32faa60f8d21"). InnerVolumeSpecName "kube-api-access-s8k2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.007945 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkcbs\" (UniqueName: \"kubernetes.io/projected/e0c0052f-9ec7-43ad-87a0-87a32f8461b7-kube-api-access-qkcbs\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.007981 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8k2l\" (UniqueName: \"kubernetes.io/projected/388347b6-ce3b-4469-ad87-32faa60f8d21-kube-api-access-s8k2l\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.007991 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0c0052f-9ec7-43ad-87a0-87a32f8461b7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.008001 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/388347b6-ce3b-4469-ad87-32faa60f8d21-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.325614 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wpt88" event={"ID":"388347b6-ce3b-4469-ad87-32faa60f8d21","Type":"ContainerDied","Data":"78715f476929389945cb38a8a8c6e374b065604d8b4e8b8d8d323f731944baa7"} Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.325638 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wpt88" Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.325652 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78715f476929389945cb38a8a8c6e374b065604d8b4e8b8d8d323f731944baa7" Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.327218 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d39d-account-create-update-hjjxx" event={"ID":"e0c0052f-9ec7-43ad-87a0-87a32f8461b7","Type":"ContainerDied","Data":"de34c4e970cb1f5de61c50ea1199585e408dc2585cdeb4f4069e394b2946f8da"} Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.327238 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de34c4e970cb1f5de61c50ea1199585e408dc2585cdeb4f4069e394b2946f8da" Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.327303 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d39d-account-create-update-hjjxx" Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.328567 4818 generic.go:334] "Generic (PLEG): container finished" podID="4803b078-4593-4b55-8a42-f4c3f4b0e21f" containerID="0e0da409a630e9d2e62668e45773b61af258dc6e15a86d708d0cad1f31f007ac" exitCode=0 Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.328678 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" event={"ID":"4803b078-4593-4b55-8a42-f4c3f4b0e21f","Type":"ContainerDied","Data":"0e0da409a630e9d2e62668e45773b61af258dc6e15a86d708d0cad1f31f007ac"} Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.328744 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" event={"ID":"4803b078-4593-4b55-8a42-f4c3f4b0e21f","Type":"ContainerStarted","Data":"99a4375a31a3b06f220595540cb9121692a2490d26e646ddea9c1148f74e88f1"} Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.817751 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0a8a-account-create-update-82q4f" Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.920162 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d32631e-bf3a-46d9-bfd3-2ea9e21987c0-operator-scripts\") pod \"4d32631e-bf3a-46d9-bfd3-2ea9e21987c0\" (UID: \"4d32631e-bf3a-46d9-bfd3-2ea9e21987c0\") " Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.920204 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h78lq\" (UniqueName: \"kubernetes.io/projected/4d32631e-bf3a-46d9-bfd3-2ea9e21987c0-kube-api-access-h78lq\") pod \"4d32631e-bf3a-46d9-bfd3-2ea9e21987c0\" (UID: \"4d32631e-bf3a-46d9-bfd3-2ea9e21987c0\") " Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.920681 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d32631e-bf3a-46d9-bfd3-2ea9e21987c0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4d32631e-bf3a-46d9-bfd3-2ea9e21987c0" (UID: "4d32631e-bf3a-46d9-bfd3-2ea9e21987c0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.920928 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d32631e-bf3a-46d9-bfd3-2ea9e21987c0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.929599 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d7v9h" Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.931865 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d32631e-bf3a-46d9-bfd3-2ea9e21987c0-kube-api-access-h78lq" (OuterVolumeSpecName: "kube-api-access-h78lq") pod "4d32631e-bf3a-46d9-bfd3-2ea9e21987c0" (UID: "4d32631e-bf3a-46d9-bfd3-2ea9e21987c0"). InnerVolumeSpecName "kube-api-access-h78lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.937647 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pdz8d" Dec 03 06:45:50 crc kubenswrapper[4818]: I1203 06:45:50.937952 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-339d-account-create-update-5hpch" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.022271 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/038a0b6a-32a3-407f-9e31-0f133592cc57-operator-scripts\") pod \"038a0b6a-32a3-407f-9e31-0f133592cc57\" (UID: \"038a0b6a-32a3-407f-9e31-0f133592cc57\") " Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.022344 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fb6x6\" (UniqueName: \"kubernetes.io/projected/939f2f3e-bc03-445c-bc1e-2cb1ca895795-kube-api-access-fb6x6\") pod \"939f2f3e-bc03-445c-bc1e-2cb1ca895795\" (UID: \"939f2f3e-bc03-445c-bc1e-2cb1ca895795\") " Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.022383 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/939f2f3e-bc03-445c-bc1e-2cb1ca895795-operator-scripts\") pod \"939f2f3e-bc03-445c-bc1e-2cb1ca895795\" (UID: \"939f2f3e-bc03-445c-bc1e-2cb1ca895795\") " Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.022567 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlx9w\" (UniqueName: \"kubernetes.io/projected/038a0b6a-32a3-407f-9e31-0f133592cc57-kube-api-access-dlx9w\") pod \"038a0b6a-32a3-407f-9e31-0f133592cc57\" (UID: \"038a0b6a-32a3-407f-9e31-0f133592cc57\") " Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.022666 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqk5c\" (UniqueName: \"kubernetes.io/projected/c8063467-b639-4aae-8d0d-db6660dfadac-kube-api-access-sqk5c\") pod \"c8063467-b639-4aae-8d0d-db6660dfadac\" (UID: \"c8063467-b639-4aae-8d0d-db6660dfadac\") " Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.022769 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8063467-b639-4aae-8d0d-db6660dfadac-operator-scripts\") pod \"c8063467-b639-4aae-8d0d-db6660dfadac\" (UID: \"c8063467-b639-4aae-8d0d-db6660dfadac\") " Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.022790 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/038a0b6a-32a3-407f-9e31-0f133592cc57-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "038a0b6a-32a3-407f-9e31-0f133592cc57" (UID: "038a0b6a-32a3-407f-9e31-0f133592cc57"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.023012 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/939f2f3e-bc03-445c-bc1e-2cb1ca895795-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "939f2f3e-bc03-445c-bc1e-2cb1ca895795" (UID: "939f2f3e-bc03-445c-bc1e-2cb1ca895795"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.023406 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8063467-b639-4aae-8d0d-db6660dfadac-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c8063467-b639-4aae-8d0d-db6660dfadac" (UID: "c8063467-b639-4aae-8d0d-db6660dfadac"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.024109 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/038a0b6a-32a3-407f-9e31-0f133592cc57-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.024145 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/939f2f3e-bc03-445c-bc1e-2cb1ca895795-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.024164 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8063467-b639-4aae-8d0d-db6660dfadac-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.024183 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h78lq\" (UniqueName: \"kubernetes.io/projected/4d32631e-bf3a-46d9-bfd3-2ea9e21987c0-kube-api-access-h78lq\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.025703 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/939f2f3e-bc03-445c-bc1e-2cb1ca895795-kube-api-access-fb6x6" (OuterVolumeSpecName: "kube-api-access-fb6x6") pod "939f2f3e-bc03-445c-bc1e-2cb1ca895795" (UID: "939f2f3e-bc03-445c-bc1e-2cb1ca895795"). InnerVolumeSpecName "kube-api-access-fb6x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.025735 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8063467-b639-4aae-8d0d-db6660dfadac-kube-api-access-sqk5c" (OuterVolumeSpecName: "kube-api-access-sqk5c") pod "c8063467-b639-4aae-8d0d-db6660dfadac" (UID: "c8063467-b639-4aae-8d0d-db6660dfadac"). InnerVolumeSpecName "kube-api-access-sqk5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.026907 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/038a0b6a-32a3-407f-9e31-0f133592cc57-kube-api-access-dlx9w" (OuterVolumeSpecName: "kube-api-access-dlx9w") pod "038a0b6a-32a3-407f-9e31-0f133592cc57" (UID: "038a0b6a-32a3-407f-9e31-0f133592cc57"). InnerVolumeSpecName "kube-api-access-dlx9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.125959 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fb6x6\" (UniqueName: \"kubernetes.io/projected/939f2f3e-bc03-445c-bc1e-2cb1ca895795-kube-api-access-fb6x6\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.126008 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlx9w\" (UniqueName: \"kubernetes.io/projected/038a0b6a-32a3-407f-9e31-0f133592cc57-kube-api-access-dlx9w\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.126023 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqk5c\" (UniqueName: \"kubernetes.io/projected/c8063467-b639-4aae-8d0d-db6660dfadac-kube-api-access-sqk5c\") on node \"crc\" DevicePath \"\"" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.336105 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pdz8d" event={"ID":"038a0b6a-32a3-407f-9e31-0f133592cc57","Type":"ContainerDied","Data":"321a89d76c01759fd51f8175ee88510ce86ce103e037f31ce238a5424439df44"} Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.336161 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="321a89d76c01759fd51f8175ee88510ce86ce103e037f31ce238a5424439df44" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.336151 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pdz8d" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.337631 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-339d-account-create-update-5hpch" event={"ID":"c8063467-b639-4aae-8d0d-db6660dfadac","Type":"ContainerDied","Data":"c581b828a2afd62a83718b90915aac028803273e9ca2efccf0639aec4218c882"} Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.337653 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c581b828a2afd62a83718b90915aac028803273e9ca2efccf0639aec4218c882" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.337702 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-339d-account-create-update-5hpch" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.343217 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-d7v9h" event={"ID":"939f2f3e-bc03-445c-bc1e-2cb1ca895795","Type":"ContainerDied","Data":"907ebae80518454f8b34ac0aad092dcb15398d322cc7159da4c3c879a0a7b2aa"} Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.343263 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="907ebae80518454f8b34ac0aad092dcb15398d322cc7159da4c3c879a0a7b2aa" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.343321 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-d7v9h" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.346131 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0a8a-account-create-update-82q4f" event={"ID":"4d32631e-bf3a-46d9-bfd3-2ea9e21987c0","Type":"ContainerDied","Data":"c4e44949771f1e325e0aa0932eb7f0afe87678b5d4d0579027ecd4846ef02617"} Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.346159 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4e44949771f1e325e0aa0932eb7f0afe87678b5d4d0579027ecd4846ef02617" Dec 03 06:45:51 crc kubenswrapper[4818]: I1203 06:45:51.346169 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0a8a-account-create-update-82q4f" Dec 03 06:45:54 crc kubenswrapper[4818]: I1203 06:45:54.376369 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" event={"ID":"4803b078-4593-4b55-8a42-f4c3f4b0e21f","Type":"ContainerStarted","Data":"7e1e0db439eb910a33d0397c6942599f54f43b4b79841c87971413ca609fdb74"} Dec 03 06:45:54 crc kubenswrapper[4818]: I1203 06:45:54.376979 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:59 crc kubenswrapper[4818]: E1203 06:45:59.035809 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e751691_c660_4527_a9a6_61505f83c6ff.slice/crio-303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0.scope\": RecentStats: unable to find data in memory cache]" Dec 03 06:45:59 crc kubenswrapper[4818]: I1203 06:45:59.084987 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:45:59 crc kubenswrapper[4818]: I1203 06:45:59.123365 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" podStartSLOduration=11.123337328 podStartE2EDuration="11.123337328s" podCreationTimestamp="2025-12-03 06:45:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:45:54.396923847 +0000 UTC m=+1112.088532599" watchObservedRunningTime="2025-12-03 06:45:59.123337328 +0000 UTC m=+1116.814946120" Dec 03 06:45:59 crc kubenswrapper[4818]: I1203 06:45:59.188692 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-8vdnp"] Dec 03 06:45:59 crc kubenswrapper[4818]: I1203 06:45:59.188977 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" podUID="b9e387e8-8f19-4d97-834d-639710ce55dc" containerName="dnsmasq-dns" containerID="cri-o://ffdc02851a91a239abe350b548ba5e893811da1284cbdc8d92949e14699a1b19" gracePeriod=10 Dec 03 06:45:59 crc kubenswrapper[4818]: I1203 06:45:59.597662 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-p6k8k" event={"ID":"ed38224f-53f7-4e63-8841-4d1c3f65c456","Type":"ContainerStarted","Data":"8805d0d2e882a8ac0a8386c703ab84d3e971b8b279955254aac512e5a53e5d58"} Dec 03 06:45:59 crc kubenswrapper[4818]: I1203 06:45:59.601186 4818 generic.go:334] "Generic (PLEG): container finished" podID="b9e387e8-8f19-4d97-834d-639710ce55dc" containerID="ffdc02851a91a239abe350b548ba5e893811da1284cbdc8d92949e14699a1b19" exitCode=0 Dec 03 06:45:59 crc kubenswrapper[4818]: I1203 06:45:59.601243 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" event={"ID":"b9e387e8-8f19-4d97-834d-639710ce55dc","Type":"ContainerDied","Data":"ffdc02851a91a239abe350b548ba5e893811da1284cbdc8d92949e14699a1b19"} Dec 03 06:45:59 crc kubenswrapper[4818]: I1203 06:45:59.620570 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-p6k8k" podStartSLOduration=3.018469562 podStartE2EDuration="13.620552494s" podCreationTimestamp="2025-12-03 06:45:46 +0000 UTC" firstStartedPulling="2025-12-03 06:45:47.847724553 +0000 UTC m=+1105.539333295" lastFinishedPulling="2025-12-03 06:45:58.449807435 +0000 UTC m=+1116.141416227" observedRunningTime="2025-12-03 06:45:59.617341134 +0000 UTC m=+1117.308949896" watchObservedRunningTime="2025-12-03 06:45:59.620552494 +0000 UTC m=+1117.312161256" Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.349432 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.518534 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-dns-swift-storage-0\") pod \"b9e387e8-8f19-4d97-834d-639710ce55dc\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.519047 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-config\") pod \"b9e387e8-8f19-4d97-834d-639710ce55dc\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.519218 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-ovsdbserver-sb\") pod \"b9e387e8-8f19-4d97-834d-639710ce55dc\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.519316 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-ovsdbserver-nb\") pod \"b9e387e8-8f19-4d97-834d-639710ce55dc\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.519382 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xml2m\" (UniqueName: \"kubernetes.io/projected/b9e387e8-8f19-4d97-834d-639710ce55dc-kube-api-access-xml2m\") pod \"b9e387e8-8f19-4d97-834d-639710ce55dc\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.519554 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-dns-svc\") pod \"b9e387e8-8f19-4d97-834d-639710ce55dc\" (UID: \"b9e387e8-8f19-4d97-834d-639710ce55dc\") " Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.526046 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9e387e8-8f19-4d97-834d-639710ce55dc-kube-api-access-xml2m" (OuterVolumeSpecName: "kube-api-access-xml2m") pod "b9e387e8-8f19-4d97-834d-639710ce55dc" (UID: "b9e387e8-8f19-4d97-834d-639710ce55dc"). InnerVolumeSpecName "kube-api-access-xml2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.564620 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b9e387e8-8f19-4d97-834d-639710ce55dc" (UID: "b9e387e8-8f19-4d97-834d-639710ce55dc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.566891 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b9e387e8-8f19-4d97-834d-639710ce55dc" (UID: "b9e387e8-8f19-4d97-834d-639710ce55dc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.582443 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b9e387e8-8f19-4d97-834d-639710ce55dc" (UID: "b9e387e8-8f19-4d97-834d-639710ce55dc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.582955 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-config" (OuterVolumeSpecName: "config") pod "b9e387e8-8f19-4d97-834d-639710ce55dc" (UID: "b9e387e8-8f19-4d97-834d-639710ce55dc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.585680 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b9e387e8-8f19-4d97-834d-639710ce55dc" (UID: "b9e387e8-8f19-4d97-834d-639710ce55dc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.610435 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" event={"ID":"b9e387e8-8f19-4d97-834d-639710ce55dc","Type":"ContainerDied","Data":"b6481b5ccf209098ca0ff0be372455fd1bcc7cd4060a6c33b396bcb80d18d02a"} Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.610485 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-8vdnp" Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.610514 4818 scope.go:117] "RemoveContainer" containerID="ffdc02851a91a239abe350b548ba5e893811da1284cbdc8d92949e14699a1b19" Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.622940 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.622994 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.623023 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.623040 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xml2m\" (UniqueName: \"kubernetes.io/projected/b9e387e8-8f19-4d97-834d-639710ce55dc-kube-api-access-xml2m\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.623056 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.623071 4818 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b9e387e8-8f19-4d97-834d-639710ce55dc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.653619 4818 scope.go:117] "RemoveContainer" containerID="e6ffe82240f0f0b63dd24e06c01f648b3492d61341cacdb4e187510f33f77bcf" Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.664196 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-8vdnp"] Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.672021 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-8vdnp"] Dec 03 06:46:00 crc kubenswrapper[4818]: I1203 06:46:00.749622 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9e387e8-8f19-4d97-834d-639710ce55dc" path="/var/lib/kubelet/pods/b9e387e8-8f19-4d97-834d-639710ce55dc/volumes" Dec 03 06:46:02 crc kubenswrapper[4818]: I1203 06:46:02.631327 4818 generic.go:334] "Generic (PLEG): container finished" podID="ed38224f-53f7-4e63-8841-4d1c3f65c456" containerID="8805d0d2e882a8ac0a8386c703ab84d3e971b8b279955254aac512e5a53e5d58" exitCode=0 Dec 03 06:46:02 crc kubenswrapper[4818]: I1203 06:46:02.631439 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-p6k8k" event={"ID":"ed38224f-53f7-4e63-8841-4d1c3f65c456","Type":"ContainerDied","Data":"8805d0d2e882a8ac0a8386c703ab84d3e971b8b279955254aac512e5a53e5d58"} Dec 03 06:46:03 crc kubenswrapper[4818]: I1203 06:46:03.973277 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-p6k8k" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.093406 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed38224f-53f7-4e63-8841-4d1c3f65c456-config-data\") pod \"ed38224f-53f7-4e63-8841-4d1c3f65c456\" (UID: \"ed38224f-53f7-4e63-8841-4d1c3f65c456\") " Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.093470 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed38224f-53f7-4e63-8841-4d1c3f65c456-combined-ca-bundle\") pod \"ed38224f-53f7-4e63-8841-4d1c3f65c456\" (UID: \"ed38224f-53f7-4e63-8841-4d1c3f65c456\") " Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.093517 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hf55t\" (UniqueName: \"kubernetes.io/projected/ed38224f-53f7-4e63-8841-4d1c3f65c456-kube-api-access-hf55t\") pod \"ed38224f-53f7-4e63-8841-4d1c3f65c456\" (UID: \"ed38224f-53f7-4e63-8841-4d1c3f65c456\") " Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.102370 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed38224f-53f7-4e63-8841-4d1c3f65c456-kube-api-access-hf55t" (OuterVolumeSpecName: "kube-api-access-hf55t") pod "ed38224f-53f7-4e63-8841-4d1c3f65c456" (UID: "ed38224f-53f7-4e63-8841-4d1c3f65c456"). InnerVolumeSpecName "kube-api-access-hf55t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.118700 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed38224f-53f7-4e63-8841-4d1c3f65c456-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed38224f-53f7-4e63-8841-4d1c3f65c456" (UID: "ed38224f-53f7-4e63-8841-4d1c3f65c456"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.137124 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed38224f-53f7-4e63-8841-4d1c3f65c456-config-data" (OuterVolumeSpecName: "config-data") pod "ed38224f-53f7-4e63-8841-4d1c3f65c456" (UID: "ed38224f-53f7-4e63-8841-4d1c3f65c456"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.195598 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed38224f-53f7-4e63-8841-4d1c3f65c456-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.195852 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed38224f-53f7-4e63-8841-4d1c3f65c456-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.195949 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hf55t\" (UniqueName: \"kubernetes.io/projected/ed38224f-53f7-4e63-8841-4d1c3f65c456-kube-api-access-hf55t\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.649966 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-p6k8k" event={"ID":"ed38224f-53f7-4e63-8841-4d1c3f65c456","Type":"ContainerDied","Data":"475139a7015ad51e6215c4bb1e1f0012056404367127ec7add4177f3e3f99d4b"} Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.650003 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="475139a7015ad51e6215c4bb1e1f0012056404367127ec7add4177f3e3f99d4b" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.650093 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-p6k8k" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.981768 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-4jjs9"] Dec 03 06:46:04 crc kubenswrapper[4818]: E1203 06:46:04.982662 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0c0052f-9ec7-43ad-87a0-87a32f8461b7" containerName="mariadb-account-create-update" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.982680 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0c0052f-9ec7-43ad-87a0-87a32f8461b7" containerName="mariadb-account-create-update" Dec 03 06:46:04 crc kubenswrapper[4818]: E1203 06:46:04.982698 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8063467-b639-4aae-8d0d-db6660dfadac" containerName="mariadb-account-create-update" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.982707 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8063467-b639-4aae-8d0d-db6660dfadac" containerName="mariadb-account-create-update" Dec 03 06:46:04 crc kubenswrapper[4818]: E1203 06:46:04.982721 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9e387e8-8f19-4d97-834d-639710ce55dc" containerName="init" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.982730 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9e387e8-8f19-4d97-834d-639710ce55dc" containerName="init" Dec 03 06:46:04 crc kubenswrapper[4818]: E1203 06:46:04.982744 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9e387e8-8f19-4d97-834d-639710ce55dc" containerName="dnsmasq-dns" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.982751 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9e387e8-8f19-4d97-834d-639710ce55dc" containerName="dnsmasq-dns" Dec 03 06:46:04 crc kubenswrapper[4818]: E1203 06:46:04.982767 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed38224f-53f7-4e63-8841-4d1c3f65c456" containerName="keystone-db-sync" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.982774 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed38224f-53f7-4e63-8841-4d1c3f65c456" containerName="keystone-db-sync" Dec 03 06:46:04 crc kubenswrapper[4818]: E1203 06:46:04.982783 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="038a0b6a-32a3-407f-9e31-0f133592cc57" containerName="mariadb-database-create" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.982791 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="038a0b6a-32a3-407f-9e31-0f133592cc57" containerName="mariadb-database-create" Dec 03 06:46:04 crc kubenswrapper[4818]: E1203 06:46:04.982800 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="388347b6-ce3b-4469-ad87-32faa60f8d21" containerName="mariadb-database-create" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.982808 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="388347b6-ce3b-4469-ad87-32faa60f8d21" containerName="mariadb-database-create" Dec 03 06:46:04 crc kubenswrapper[4818]: E1203 06:46:04.982850 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d32631e-bf3a-46d9-bfd3-2ea9e21987c0" containerName="mariadb-account-create-update" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.982859 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d32631e-bf3a-46d9-bfd3-2ea9e21987c0" containerName="mariadb-account-create-update" Dec 03 06:46:04 crc kubenswrapper[4818]: E1203 06:46:04.982883 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="939f2f3e-bc03-445c-bc1e-2cb1ca895795" containerName="mariadb-database-create" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.982892 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="939f2f3e-bc03-445c-bc1e-2cb1ca895795" containerName="mariadb-database-create" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.983089 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed38224f-53f7-4e63-8841-4d1c3f65c456" containerName="keystone-db-sync" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.983108 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8063467-b639-4aae-8d0d-db6660dfadac" containerName="mariadb-account-create-update" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.983126 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9e387e8-8f19-4d97-834d-639710ce55dc" containerName="dnsmasq-dns" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.983143 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="939f2f3e-bc03-445c-bc1e-2cb1ca895795" containerName="mariadb-database-create" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.983160 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="038a0b6a-32a3-407f-9e31-0f133592cc57" containerName="mariadb-database-create" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.983169 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="388347b6-ce3b-4469-ad87-32faa60f8d21" containerName="mariadb-database-create" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.983182 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0c0052f-9ec7-43ad-87a0-87a32f8461b7" containerName="mariadb-account-create-update" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.983192 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d32631e-bf3a-46d9-bfd3-2ea9e21987c0" containerName="mariadb-account-create-update" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.988399 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.991090 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.991173 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.991281 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ggd6v" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.994786 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 06:46:04 crc kubenswrapper[4818]: I1203 06:46:04.994847 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.005873 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4jjs9"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.011882 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-fernet-keys\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.011970 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-credential-keys\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.012033 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-combined-ca-bundle\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.012089 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-config-data\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.012109 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pktv2\" (UniqueName: \"kubernetes.io/projected/42277c61-53cd-4617-b5bd-ba45863243d2-kube-api-access-pktv2\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.012296 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-scripts\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.060391 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-c6nvr"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.062651 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.071638 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-c6nvr"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.116726 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-config\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.116835 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-fernet-keys\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.116876 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-credential-keys\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.116910 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-combined-ca-bundle\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.116955 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-config-data\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.116983 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pktv2\" (UniqueName: \"kubernetes.io/projected/42277c61-53cd-4617-b5bd-ba45863243d2-kube-api-access-pktv2\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.117037 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-dns-svc\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.117061 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.117089 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwbrv\" (UniqueName: \"kubernetes.io/projected/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-kube-api-access-kwbrv\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.117119 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-scripts\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.117145 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.117169 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.134850 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-scripts\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.138370 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-fernet-keys\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.138592 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-credential-keys\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.138882 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-combined-ca-bundle\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.141620 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-config-data\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.186590 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pktv2\" (UniqueName: \"kubernetes.io/projected/42277c61-53cd-4617-b5bd-ba45863243d2-kube-api-access-pktv2\") pod \"keystone-bootstrap-4jjs9\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.208231 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5649b85449-wr6w6"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.214296 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.218566 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.218624 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwbrv\" (UniqueName: \"kubernetes.io/projected/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-kube-api-access-kwbrv\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.218655 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.218677 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.218721 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-horizon-secret-key\") pod \"horizon-5649b85449-wr6w6\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.218756 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-scripts\") pod \"horizon-5649b85449-wr6w6\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.218781 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-config-data\") pod \"horizon-5649b85449-wr6w6\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.218831 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-config\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.218890 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgpl2\" (UniqueName: \"kubernetes.io/projected/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-kube-api-access-xgpl2\") pod \"horizon-5649b85449-wr6w6\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.218934 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-dns-svc\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.218954 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-logs\") pod \"horizon-5649b85449-wr6w6\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.220102 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.220592 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-config\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.221013 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.221064 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.225457 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.225613 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-r9n6p" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.225698 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.225759 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.226791 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-dns-svc\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.248966 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-978k5"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.251354 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.266006 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-m2q9c" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.266672 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.281796 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.284242 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwbrv\" (UniqueName: \"kubernetes.io/projected/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-kube-api-access-kwbrv\") pod \"dnsmasq-dns-847c4cc679-c6nvr\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.285201 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5649b85449-wr6w6"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.321244 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.325915 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-config-data\") pod \"horizon-5649b85449-wr6w6\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.326606 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgpl2\" (UniqueName: \"kubernetes.io/projected/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-kube-api-access-xgpl2\") pod \"horizon-5649b85449-wr6w6\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.326782 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-logs\") pod \"horizon-5649b85449-wr6w6\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.326941 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-horizon-secret-key\") pod \"horizon-5649b85449-wr6w6\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.327058 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-scripts\") pod \"horizon-5649b85449-wr6w6\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.331717 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-scripts\") pod \"horizon-5649b85449-wr6w6\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.333795 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-978k5"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.341073 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-config-data\") pod \"horizon-5649b85449-wr6w6\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.341433 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-logs\") pod \"horizon-5649b85449-wr6w6\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.355221 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-horizon-secret-key\") pod \"horizon-5649b85449-wr6w6\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.377400 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgpl2\" (UniqueName: \"kubernetes.io/projected/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-kube-api-access-xgpl2\") pod \"horizon-5649b85449-wr6w6\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.400848 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.416796 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-ffzld"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.421536 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-ffzld" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.424165 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-zctkr" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.424373 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.426160 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.428727 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-scripts\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.428779 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhh5p\" (UniqueName: \"kubernetes.io/projected/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-kube-api-access-xhh5p\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.428806 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-db-sync-config-data\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.428885 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-etc-machine-id\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.428908 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-config-data\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.428946 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-combined-ca-bundle\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.452585 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.464339 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.478278 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.478475 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.482266 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-ffzld"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.531386 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-etc-machine-id\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.531461 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-config-data\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.531505 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-combined-ca-bundle\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.531554 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv774\" (UniqueName: \"kubernetes.io/projected/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-kube-api-access-dv774\") pod \"neutron-db-sync-ffzld\" (UID: \"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7\") " pod="openstack/neutron-db-sync-ffzld" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.531633 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-scripts\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.531661 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhh5p\" (UniqueName: \"kubernetes.io/projected/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-kube-api-access-xhh5p\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.531695 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-db-sync-config-data\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.531711 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-config\") pod \"neutron-db-sync-ffzld\" (UID: \"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7\") " pod="openstack/neutron-db-sync-ffzld" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.531774 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-combined-ca-bundle\") pod \"neutron-db-sync-ffzld\" (UID: \"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7\") " pod="openstack/neutron-db-sync-ffzld" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.531926 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-etc-machine-id\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.568780 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-scripts\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.569925 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-combined-ca-bundle\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.585102 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-config-data\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.629124 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.630402 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-db-sync-config-data\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.639967 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.640035 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/575fe180-2a23-48e9-b5cf-b90e83528e5e-log-httpd\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.640101 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-config\") pod \"neutron-db-sync-ffzld\" (UID: \"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7\") " pod="openstack/neutron-db-sync-ffzld" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.640186 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-config-data\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.640222 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-combined-ca-bundle\") pod \"neutron-db-sync-ffzld\" (UID: \"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7\") " pod="openstack/neutron-db-sync-ffzld" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.640254 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-scripts\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.640298 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d2sg\" (UniqueName: \"kubernetes.io/projected/575fe180-2a23-48e9-b5cf-b90e83528e5e-kube-api-access-5d2sg\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.640338 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/575fe180-2a23-48e9-b5cf-b90e83528e5e-run-httpd\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.640368 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.640444 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv774\" (UniqueName: \"kubernetes.io/projected/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-kube-api-access-dv774\") pod \"neutron-db-sync-ffzld\" (UID: \"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7\") " pod="openstack/neutron-db-sync-ffzld" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.679471 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhh5p\" (UniqueName: \"kubernetes.io/projected/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-kube-api-access-xhh5p\") pod \"cinder-db-sync-978k5\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.688189 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-978k5" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.705281 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv774\" (UniqueName: \"kubernetes.io/projected/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-kube-api-access-dv774\") pod \"neutron-db-sync-ffzld\" (UID: \"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7\") " pod="openstack/neutron-db-sync-ffzld" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.706105 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-config\") pod \"neutron-db-sync-ffzld\" (UID: \"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7\") " pod="openstack/neutron-db-sync-ffzld" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.722079 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-combined-ca-bundle\") pod \"neutron-db-sync-ffzld\" (UID: \"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7\") " pod="openstack/neutron-db-sync-ffzld" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.736924 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.741944 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.741978 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/575fe180-2a23-48e9-b5cf-b90e83528e5e-log-httpd\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.742032 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-config-data\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.742057 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-scripts\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.742080 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d2sg\" (UniqueName: \"kubernetes.io/projected/575fe180-2a23-48e9-b5cf-b90e83528e5e-kube-api-access-5d2sg\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.742107 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/575fe180-2a23-48e9-b5cf-b90e83528e5e-run-httpd\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.745740 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/575fe180-2a23-48e9-b5cf-b90e83528e5e-run-httpd\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.747110 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.748039 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/575fe180-2a23-48e9-b5cf-b90e83528e5e-log-httpd\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.753643 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-ffzld" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.754328 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-scripts\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.758001 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.772061 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.782183 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d2sg\" (UniqueName: \"kubernetes.io/projected/575fe180-2a23-48e9-b5cf-b90e83528e5e-kube-api-access-5d2sg\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.785696 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-config-data\") pod \"ceilometer-0\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.802949 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7fbcd7d5fc-6stq9"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.804662 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.814475 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.834398 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-lqzrp"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.836295 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.839887 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.840067 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-n4gmb" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.840223 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.847391 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.849636 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.855713 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.855970 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.856081 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-zqbs6" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.857867 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.860063 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-c6nvr"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.877661 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-lqzrp"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.890567 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.909945 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7fbcd7d5fc-6stq9"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.920518 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-7kbvz"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.922623 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7kbvz" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.927661 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-2n894" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.927699 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.940211 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-7kbvz"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.949425 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-9kv6h"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952459 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9d0cba79-7a13-40ee-ab38-ae093b204a9c-horizon-secret-key\") pod \"horizon-7fbcd7d5fc-6stq9\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952513 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952548 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c23baebd-4d73-4972-b9eb-5c573e859977-logs\") pod \"placement-db-sync-lqzrp\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952565 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952593 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d0cba79-7a13-40ee-ab38-ae093b204a9c-config-data\") pod \"horizon-7fbcd7d5fc-6stq9\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952621 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp7qh\" (UniqueName: \"kubernetes.io/projected/9d0cba79-7a13-40ee-ab38-ae093b204a9c-kube-api-access-xp7qh\") pod \"horizon-7fbcd7d5fc-6stq9\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952657 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952681 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-config-data\") pod \"placement-db-sync-lqzrp\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952734 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952759 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-scripts\") pod \"placement-db-sync-lqzrp\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952784 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d0cba79-7a13-40ee-ab38-ae093b204a9c-logs\") pod \"horizon-7fbcd7d5fc-6stq9\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952802 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zdmf\" (UniqueName: \"kubernetes.io/projected/c23baebd-4d73-4972-b9eb-5c573e859977-kube-api-access-8zdmf\") pod \"placement-db-sync-lqzrp\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952834 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2af4147-3de6-499c-a6f3-133b3e16d673-logs\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952855 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2af4147-3de6-499c-a6f3-133b3e16d673-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952885 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d0cba79-7a13-40ee-ab38-ae093b204a9c-scripts\") pod \"horizon-7fbcd7d5fc-6stq9\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952902 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-combined-ca-bundle\") pod \"placement-db-sync-lqzrp\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952931 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.952954 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9kh5\" (UniqueName: \"kubernetes.io/projected/a2af4147-3de6-499c-a6f3-133b3e16d673-kube-api-access-d9kh5\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.955644 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.962016 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-9kv6h"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.973710 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.979694 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.981937 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.982235 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 06:46:05 crc kubenswrapper[4818]: I1203 06:46:05.985605 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.057648 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.057704 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d0cba79-7a13-40ee-ab38-ae093b204a9c-config-data\") pod \"horizon-7fbcd7d5fc-6stq9\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.057738 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp7qh\" (UniqueName: \"kubernetes.io/projected/9d0cba79-7a13-40ee-ab38-ae093b204a9c-kube-api-access-xp7qh\") pod \"horizon-7fbcd7d5fc-6stq9\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.057764 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.057786 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.057807 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-config-data\") pod \"placement-db-sync-lqzrp\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.057920 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.057943 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-scripts\") pod \"placement-db-sync-lqzrp\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.057965 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d0cba79-7a13-40ee-ab38-ae093b204a9c-logs\") pod \"horizon-7fbcd7d5fc-6stq9\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.057982 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zdmf\" (UniqueName: \"kubernetes.io/projected/c23baebd-4d73-4972-b9eb-5c573e859977-kube-api-access-8zdmf\") pod \"placement-db-sync-lqzrp\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.058003 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-config\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.058018 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2af4147-3de6-499c-a6f3-133b3e16d673-logs\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.058033 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2af4147-3de6-499c-a6f3-133b3e16d673-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.058063 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d0cba79-7a13-40ee-ab38-ae093b204a9c-scripts\") pod \"horizon-7fbcd7d5fc-6stq9\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.058082 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-combined-ca-bundle\") pod \"placement-db-sync-lqzrp\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.058102 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.058124 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.058145 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9kh5\" (UniqueName: \"kubernetes.io/projected/a2af4147-3de6-499c-a6f3-133b3e16d673-kube-api-access-d9kh5\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.058162 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct4xm\" (UniqueName: \"kubernetes.io/projected/54cbca7f-1646-4eaa-a61b-6094b8506572-kube-api-access-ct4xm\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.058183 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.058218 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9d0cba79-7a13-40ee-ab38-ae093b204a9c-horizon-secret-key\") pod \"horizon-7fbcd7d5fc-6stq9\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.058240 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.058266 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c23baebd-4d73-4972-b9eb-5c573e859977-logs\") pod \"placement-db-sync-lqzrp\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.058282 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.059585 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2af4147-3de6-499c-a6f3-133b3e16d673-logs\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.060003 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2af4147-3de6-499c-a6f3-133b3e16d673-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.060757 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d0cba79-7a13-40ee-ab38-ae093b204a9c-config-data\") pod \"horizon-7fbcd7d5fc-6stq9\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.062319 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d0cba79-7a13-40ee-ab38-ae093b204a9c-scripts\") pod \"horizon-7fbcd7d5fc-6stq9\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.066711 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-scripts\") pod \"placement-db-sync-lqzrp\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.066981 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-combined-ca-bundle\") pod \"placement-db-sync-lqzrp\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.067699 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.068321 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9d0cba79-7a13-40ee-ab38-ae093b204a9c-horizon-secret-key\") pod \"horizon-7fbcd7d5fc-6stq9\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.068384 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c23baebd-4d73-4972-b9eb-5c573e859977-logs\") pod \"placement-db-sync-lqzrp\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.068556 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d0cba79-7a13-40ee-ab38-ae093b204a9c-logs\") pod \"horizon-7fbcd7d5fc-6stq9\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.070602 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.070799 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-config-data\") pod \"placement-db-sync-lqzrp\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.071493 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.071591 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.074886 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.090174 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp7qh\" (UniqueName: \"kubernetes.io/projected/9d0cba79-7a13-40ee-ab38-ae093b204a9c-kube-api-access-xp7qh\") pod \"horizon-7fbcd7d5fc-6stq9\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.094466 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9kh5\" (UniqueName: \"kubernetes.io/projected/a2af4147-3de6-499c-a6f3-133b3e16d673-kube-api-access-d9kh5\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.104327 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zdmf\" (UniqueName: \"kubernetes.io/projected/c23baebd-4d73-4972-b9eb-5c573e859977-kube-api-access-8zdmf\") pod \"placement-db-sync-lqzrp\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.119705 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.126591 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4jjs9"] Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.200355 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.200416 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.201271 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.202091 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/982110da-ac09-4df8-9729-77284ed7539b-db-sync-config-data\") pod \"barbican-db-sync-7kbvz\" (UID: \"982110da-ac09-4df8-9729-77284ed7539b\") " pod="openstack/barbican-db-sync-7kbvz" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.202140 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.202168 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c50d74e9-8f1f-43f7-ad45-3a9129974cee-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.202200 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.202223 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c50d74e9-8f1f-43f7-ad45-3a9129974cee-logs\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.202259 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.202281 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.203091 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.203580 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.219167 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-scripts\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.219238 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-config-data\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.219316 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/982110da-ac09-4df8-9729-77284ed7539b-combined-ca-bundle\") pod \"barbican-db-sync-7kbvz\" (UID: \"982110da-ac09-4df8-9729-77284ed7539b\") " pod="openstack/barbican-db-sync-7kbvz" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.219354 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-config\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.219404 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79bkg\" (UniqueName: \"kubernetes.io/projected/c50d74e9-8f1f-43f7-ad45-3a9129974cee-kube-api-access-79bkg\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.219450 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.219470 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.219524 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct4xm\" (UniqueName: \"kubernetes.io/projected/54cbca7f-1646-4eaa-a61b-6094b8506572-kube-api-access-ct4xm\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.219556 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.219578 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwqv8\" (UniqueName: \"kubernetes.io/projected/982110da-ac09-4df8-9729-77284ed7539b-kube-api-access-vwqv8\") pod \"barbican-db-sync-7kbvz\" (UID: \"982110da-ac09-4df8-9729-77284ed7539b\") " pod="openstack/barbican-db-sync-7kbvz" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.220415 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-config\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.220648 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.221386 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.239358 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct4xm\" (UniqueName: \"kubernetes.io/projected/54cbca7f-1646-4eaa-a61b-6094b8506572-kube-api-access-ct4xm\") pod \"dnsmasq-dns-785d8bcb8c-9kv6h\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.300702 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.322801 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-config-data\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.322874 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/982110da-ac09-4df8-9729-77284ed7539b-combined-ca-bundle\") pod \"barbican-db-sync-7kbvz\" (UID: \"982110da-ac09-4df8-9729-77284ed7539b\") " pod="openstack/barbican-db-sync-7kbvz" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.322908 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79bkg\" (UniqueName: \"kubernetes.io/projected/c50d74e9-8f1f-43f7-ad45-3a9129974cee-kube-api-access-79bkg\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.322936 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.322973 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwqv8\" (UniqueName: \"kubernetes.io/projected/982110da-ac09-4df8-9729-77284ed7539b-kube-api-access-vwqv8\") pod \"barbican-db-sync-7kbvz\" (UID: \"982110da-ac09-4df8-9729-77284ed7539b\") " pod="openstack/barbican-db-sync-7kbvz" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.323013 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/982110da-ac09-4df8-9729-77284ed7539b-db-sync-config-data\") pod \"barbican-db-sync-7kbvz\" (UID: \"982110da-ac09-4df8-9729-77284ed7539b\") " pod="openstack/barbican-db-sync-7kbvz" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.323029 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c50d74e9-8f1f-43f7-ad45-3a9129974cee-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.323051 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.323072 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c50d74e9-8f1f-43f7-ad45-3a9129974cee-logs\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.323106 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.323133 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-scripts\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.324462 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c50d74e9-8f1f-43f7-ad45-3a9129974cee-logs\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.324993 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.325915 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c50d74e9-8f1f-43f7-ad45-3a9129974cee-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.338838 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-scripts\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.340579 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/982110da-ac09-4df8-9729-77284ed7539b-db-sync-config-data\") pod \"barbican-db-sync-7kbvz\" (UID: \"982110da-ac09-4df8-9729-77284ed7539b\") " pod="openstack/barbican-db-sync-7kbvz" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.341143 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/982110da-ac09-4df8-9729-77284ed7539b-combined-ca-bundle\") pod \"barbican-db-sync-7kbvz\" (UID: \"982110da-ac09-4df8-9729-77284ed7539b\") " pod="openstack/barbican-db-sync-7kbvz" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.346336 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.346787 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwqv8\" (UniqueName: \"kubernetes.io/projected/982110da-ac09-4df8-9729-77284ed7539b-kube-api-access-vwqv8\") pod \"barbican-db-sync-7kbvz\" (UID: \"982110da-ac09-4df8-9729-77284ed7539b\") " pod="openstack/barbican-db-sync-7kbvz" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.351517 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-c6nvr"] Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.352653 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-config-data\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.385925 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.389893 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79bkg\" (UniqueName: \"kubernetes.io/projected/c50d74e9-8f1f-43f7-ad45-3a9129974cee-kube-api-access-79bkg\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.401884 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.405670 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5649b85449-wr6w6"] Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.546673 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7kbvz" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.590904 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-ffzld"] Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.606182 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.733167 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" event={"ID":"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea","Type":"ContainerStarted","Data":"af42759ac18333729e516048febe84ad6c47de38a154565a92deae9e95ba2e46"} Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.794415 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.794458 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-978k5"] Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.794473 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4jjs9" event={"ID":"42277c61-53cd-4617-b5bd-ba45863243d2","Type":"ContainerStarted","Data":"c4dbb7ed6dec28ce96cc06f3ce4b5897d894ddbf58b71f76ad3d2c99fe295f1c"} Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.794496 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-ffzld" event={"ID":"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7","Type":"ContainerStarted","Data":"4bfa7ccda7f77d76153ede6ac6a2ccddc8f21655beef1f0b5cae25a27aa84758"} Dec 03 06:46:06 crc kubenswrapper[4818]: I1203 06:46:06.794511 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5649b85449-wr6w6" event={"ID":"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9","Type":"ContainerStarted","Data":"850ead00d645ec72cc20fd3abd75672540a69c0fe3b9f172e911ac43851dac62"} Dec 03 06:46:06 crc kubenswrapper[4818]: W1203 06:46:06.828514 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod575fe180_2a23_48e9_b5cf_b90e83528e5e.slice/crio-bc146155042cb9aec31b837131f44bfd6a7e5c6593212a004dcf42473892a271 WatchSource:0}: Error finding container bc146155042cb9aec31b837131f44bfd6a7e5c6593212a004dcf42473892a271: Status 404 returned error can't find the container with id bc146155042cb9aec31b837131f44bfd6a7e5c6593212a004dcf42473892a271 Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.007141 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7fbcd7d5fc-6stq9"] Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.354067 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-7kbvz"] Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.372060 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-lqzrp"] Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.415006 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-9kv6h"] Dec 03 06:46:07 crc kubenswrapper[4818]: W1203 06:46:07.426655 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54cbca7f_1646_4eaa_a61b_6094b8506572.slice/crio-3fa180d4d46f7e5941905882704b9e463c75c28082bfd56e0a362d8b4d8615b9 WatchSource:0}: Error finding container 3fa180d4d46f7e5941905882704b9e463c75c28082bfd56e0a362d8b4d8615b9: Status 404 returned error can't find the container with id 3fa180d4d46f7e5941905882704b9e463c75c28082bfd56e0a362d8b4d8615b9 Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.524209 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.625495 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.640186 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5649b85449-wr6w6"] Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.700953 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-797779bc9c-df4k9"] Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.703100 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.771458 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f18580e-5450-4edb-b266-734372950182-scripts\") pod \"horizon-797779bc9c-df4k9\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.771960 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f18580e-5450-4edb-b266-734372950182-logs\") pod \"horizon-797779bc9c-df4k9\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.772083 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmbqm\" (UniqueName: \"kubernetes.io/projected/2f18580e-5450-4edb-b266-734372950182-kube-api-access-vmbqm\") pod \"horizon-797779bc9c-df4k9\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.772464 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f18580e-5450-4edb-b266-734372950182-config-data\") pod \"horizon-797779bc9c-df4k9\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.772723 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2f18580e-5450-4edb-b266-734372950182-horizon-secret-key\") pod \"horizon-797779bc9c-df4k9\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.772993 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-797779bc9c-df4k9"] Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.786991 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7kbvz" event={"ID":"982110da-ac09-4df8-9729-77284ed7539b","Type":"ContainerStarted","Data":"b870c49ea2ff79b6648f1f60c7f0e0f53ec7381310232bb5cc9985375a4527ab"} Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.805147 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fbcd7d5fc-6stq9" event={"ID":"9d0cba79-7a13-40ee-ab38-ae093b204a9c","Type":"ContainerStarted","Data":"43fb75ae5439022048bfc06ddf6cb52024a6a7f9401d56d1da180212077cfd88"} Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.808576 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"575fe180-2a23-48e9-b5cf-b90e83528e5e","Type":"ContainerStarted","Data":"bc146155042cb9aec31b837131f44bfd6a7e5c6593212a004dcf42473892a271"} Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.812176 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.815731 4818 generic.go:334] "Generic (PLEG): container finished" podID="8bfb1795-6ea1-4d0d-bcd1-77937bf698ea" containerID="f5698d9a4cc501a11d2f6c99d65b0bc98bfaa94cd839ae515030ef66a6105da6" exitCode=0 Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.815983 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" event={"ID":"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea","Type":"ContainerDied","Data":"f5698d9a4cc501a11d2f6c99d65b0bc98bfaa94cd839ae515030ef66a6105da6"} Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.826224 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4jjs9" event={"ID":"42277c61-53cd-4617-b5bd-ba45863243d2","Type":"ContainerStarted","Data":"a27026bc5f78f093258fff7de41856c519084974e28212246f72c061552e254f"} Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.830745 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-978k5" event={"ID":"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3","Type":"ContainerStarted","Data":"7bb6cbbd03d8420de615001cd19501e708e819de1225b337dc628b6666a4e468"} Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.833682 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.837544 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-ffzld" event={"ID":"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7","Type":"ContainerStarted","Data":"a1d52d0e41b793fc88a51dc145961cc2a6b3b6f46b3ca4f38a0e30404c4e6cdf"} Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.840001 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2af4147-3de6-499c-a6f3-133b3e16d673","Type":"ContainerStarted","Data":"984be78d81560ce343a6cc21d1c7c233b96d9c10c2be88a9ae778816ab387230"} Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.846286 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lqzrp" event={"ID":"c23baebd-4d73-4972-b9eb-5c573e859977","Type":"ContainerStarted","Data":"e540c77a9433517242e3f821216550b8dc2f195a78e8c037fff11108cca47a4e"} Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.853110 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" event={"ID":"54cbca7f-1646-4eaa-a61b-6094b8506572","Type":"ContainerStarted","Data":"3fa180d4d46f7e5941905882704b9e463c75c28082bfd56e0a362d8b4d8615b9"} Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.857296 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.880541 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2f18580e-5450-4edb-b266-734372950182-horizon-secret-key\") pod \"horizon-797779bc9c-df4k9\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.880686 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f18580e-5450-4edb-b266-734372950182-scripts\") pod \"horizon-797779bc9c-df4k9\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.880740 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f18580e-5450-4edb-b266-734372950182-logs\") pod \"horizon-797779bc9c-df4k9\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.880755 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmbqm\" (UniqueName: \"kubernetes.io/projected/2f18580e-5450-4edb-b266-734372950182-kube-api-access-vmbqm\") pod \"horizon-797779bc9c-df4k9\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.880801 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f18580e-5450-4edb-b266-734372950182-config-data\") pod \"horizon-797779bc9c-df4k9\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.882507 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f18580e-5450-4edb-b266-734372950182-scripts\") pod \"horizon-797779bc9c-df4k9\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.883474 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f18580e-5450-4edb-b266-734372950182-logs\") pod \"horizon-797779bc9c-df4k9\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.886110 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f18580e-5450-4edb-b266-734372950182-config-data\") pod \"horizon-797779bc9c-df4k9\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.895130 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2f18580e-5450-4edb-b266-734372950182-horizon-secret-key\") pod \"horizon-797779bc9c-df4k9\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.911801 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmbqm\" (UniqueName: \"kubernetes.io/projected/2f18580e-5450-4edb-b266-734372950182-kube-api-access-vmbqm\") pod \"horizon-797779bc9c-df4k9\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.917204 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-ffzld" podStartSLOduration=2.917180628 podStartE2EDuration="2.917180628s" podCreationTimestamp="2025-12-03 06:46:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:46:07.861239205 +0000 UTC m=+1125.552847957" watchObservedRunningTime="2025-12-03 06:46:07.917180628 +0000 UTC m=+1125.608789380" Dec 03 06:46:07 crc kubenswrapper[4818]: I1203 06:46:07.918883 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-4jjs9" podStartSLOduration=3.91887426 podStartE2EDuration="3.91887426s" podCreationTimestamp="2025-12-03 06:46:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:46:07.892607 +0000 UTC m=+1125.584215752" watchObservedRunningTime="2025-12-03 06:46:07.91887426 +0000 UTC m=+1125.610483012" Dec 03 06:46:08 crc kubenswrapper[4818]: I1203 06:46:08.060296 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.428410 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.596009 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-ovsdbserver-sb\") pod \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.596100 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-dns-swift-storage-0\") pod \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.596201 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-ovsdbserver-nb\") pod \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.596255 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwbrv\" (UniqueName: \"kubernetes.io/projected/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-kube-api-access-kwbrv\") pod \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.596303 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-config\") pod \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.596324 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-dns-svc\") pod \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\" (UID: \"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea\") " Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.625734 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-kube-api-access-kwbrv" (OuterVolumeSpecName: "kube-api-access-kwbrv") pod "8bfb1795-6ea1-4d0d-bcd1-77937bf698ea" (UID: "8bfb1795-6ea1-4d0d-bcd1-77937bf698ea"). InnerVolumeSpecName "kube-api-access-kwbrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.641037 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8bfb1795-6ea1-4d0d-bcd1-77937bf698ea" (UID: "8bfb1795-6ea1-4d0d-bcd1-77937bf698ea"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.641845 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8bfb1795-6ea1-4d0d-bcd1-77937bf698ea" (UID: "8bfb1795-6ea1-4d0d-bcd1-77937bf698ea"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.646841 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8bfb1795-6ea1-4d0d-bcd1-77937bf698ea" (UID: "8bfb1795-6ea1-4d0d-bcd1-77937bf698ea"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.653787 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8bfb1795-6ea1-4d0d-bcd1-77937bf698ea" (UID: "8bfb1795-6ea1-4d0d-bcd1-77937bf698ea"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.661078 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-config" (OuterVolumeSpecName: "config") pod "8bfb1795-6ea1-4d0d-bcd1-77937bf698ea" (UID: "8bfb1795-6ea1-4d0d-bcd1-77937bf698ea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.698096 4818 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.698129 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.698141 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwbrv\" (UniqueName: \"kubernetes.io/projected/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-kube-api-access-kwbrv\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.698155 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.698167 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.698177 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.867129 4818 generic.go:334] "Generic (PLEG): container finished" podID="54cbca7f-1646-4eaa-a61b-6094b8506572" containerID="22290b7ddabe8ab3beed9e02937c9b2e354cb3bc3c99f0fcf90aa565869c3505" exitCode=0 Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.867197 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" event={"ID":"54cbca7f-1646-4eaa-a61b-6094b8506572","Type":"ContainerDied","Data":"22290b7ddabe8ab3beed9e02937c9b2e354cb3bc3c99f0fcf90aa565869c3505"} Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.872337 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" event={"ID":"8bfb1795-6ea1-4d0d-bcd1-77937bf698ea","Type":"ContainerDied","Data":"af42759ac18333729e516048febe84ad6c47de38a154565a92deae9e95ba2e46"} Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.872383 4818 scope.go:117] "RemoveContainer" containerID="f5698d9a4cc501a11d2f6c99d65b0bc98bfaa94cd839ae515030ef66a6105da6" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.872607 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-c6nvr" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.877006 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c50d74e9-8f1f-43f7-ad45-3a9129974cee","Type":"ContainerStarted","Data":"c1b13df7785d17fa5a6657f602efc94d7b5f9a6edc5bd4c712539587e65a3395"} Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.950376 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-c6nvr"] Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:08.957465 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-c6nvr"] Dec 03 06:46:09 crc kubenswrapper[4818]: E1203 06:46:09.337649 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e751691_c660_4527_a9a6_61505f83c6ff.slice/crio-303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0.scope\": RecentStats: unable to find data in memory cache]" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:09.909616 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c50d74e9-8f1f-43f7-ad45-3a9129974cee","Type":"ContainerStarted","Data":"98610b9841b1e3b474cfcdf1434dbefe1d8b882c511d2a90727870f68594b9c5"} Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:09.915204 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" event={"ID":"54cbca7f-1646-4eaa-a61b-6094b8506572","Type":"ContainerStarted","Data":"8dd11ab6b6e7e9a1512584a33ea49bca6d1ef864366536d8dc3ad47ffb615895"} Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:09.915660 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:09.922993 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2af4147-3de6-499c-a6f3-133b3e16d673","Type":"ContainerStarted","Data":"0cf9e04ee236734f6aa97650dbff35720342c9a3a05bf4c892c2cd049aca1c32"} Dec 03 06:46:09 crc kubenswrapper[4818]: I1203 06:46:09.947602 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" podStartSLOduration=4.947572804 podStartE2EDuration="4.947572804s" podCreationTimestamp="2025-12-03 06:46:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:46:09.937460234 +0000 UTC m=+1127.629068986" watchObservedRunningTime="2025-12-03 06:46:09.947572804 +0000 UTC m=+1127.639181556" Dec 03 06:46:10 crc kubenswrapper[4818]: I1203 06:46:10.129096 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-797779bc9c-df4k9"] Dec 03 06:46:10 crc kubenswrapper[4818]: I1203 06:46:10.760212 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bfb1795-6ea1-4d0d-bcd1-77937bf698ea" path="/var/lib/kubelet/pods/8bfb1795-6ea1-4d0d-bcd1-77937bf698ea/volumes" Dec 03 06:46:10 crc kubenswrapper[4818]: I1203 06:46:10.935053 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-797779bc9c-df4k9" event={"ID":"2f18580e-5450-4edb-b266-734372950182","Type":"ContainerStarted","Data":"4de6cea32430cd98c694d5818917fcda1879030caecf10d9124e3eb917a1a81b"} Dec 03 06:46:11 crc kubenswrapper[4818]: I1203 06:46:11.949333 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2af4147-3de6-499c-a6f3-133b3e16d673","Type":"ContainerStarted","Data":"fb3177c3e1157b4647a93a74fcb0e22835768f704f8b46e33abb77495a11ddc3"} Dec 03 06:46:11 crc kubenswrapper[4818]: I1203 06:46:11.949828 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a2af4147-3de6-499c-a6f3-133b3e16d673" containerName="glance-log" containerID="cri-o://0cf9e04ee236734f6aa97650dbff35720342c9a3a05bf4c892c2cd049aca1c32" gracePeriod=30 Dec 03 06:46:11 crc kubenswrapper[4818]: I1203 06:46:11.950273 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a2af4147-3de6-499c-a6f3-133b3e16d673" containerName="glance-httpd" containerID="cri-o://fb3177c3e1157b4647a93a74fcb0e22835768f704f8b46e33abb77495a11ddc3" gracePeriod=30 Dec 03 06:46:11 crc kubenswrapper[4818]: I1203 06:46:11.959980 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c50d74e9-8f1f-43f7-ad45-3a9129974cee","Type":"ContainerStarted","Data":"3d9905e829fb27f0002b396cc9c3370b19054e14f905ca2626785b32e804fc2a"} Dec 03 06:46:11 crc kubenswrapper[4818]: I1203 06:46:11.960168 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c50d74e9-8f1f-43f7-ad45-3a9129974cee" containerName="glance-log" containerID="cri-o://98610b9841b1e3b474cfcdf1434dbefe1d8b882c511d2a90727870f68594b9c5" gracePeriod=30 Dec 03 06:46:11 crc kubenswrapper[4818]: I1203 06:46:11.960207 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c50d74e9-8f1f-43f7-ad45-3a9129974cee" containerName="glance-httpd" containerID="cri-o://3d9905e829fb27f0002b396cc9c3370b19054e14f905ca2626785b32e804fc2a" gracePeriod=30 Dec 03 06:46:11 crc kubenswrapper[4818]: I1203 06:46:11.965020 4818 generic.go:334] "Generic (PLEG): container finished" podID="42277c61-53cd-4617-b5bd-ba45863243d2" containerID="a27026bc5f78f093258fff7de41856c519084974e28212246f72c061552e254f" exitCode=0 Dec 03 06:46:11 crc kubenswrapper[4818]: I1203 06:46:11.965093 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4jjs9" event={"ID":"42277c61-53cd-4617-b5bd-ba45863243d2","Type":"ContainerDied","Data":"a27026bc5f78f093258fff7de41856c519084974e28212246f72c061552e254f"} Dec 03 06:46:11 crc kubenswrapper[4818]: I1203 06:46:11.977653 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.977634632 podStartE2EDuration="6.977634632s" podCreationTimestamp="2025-12-03 06:46:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:46:11.973357347 +0000 UTC m=+1129.664966109" watchObservedRunningTime="2025-12-03 06:46:11.977634632 +0000 UTC m=+1129.669243384" Dec 03 06:46:12 crc kubenswrapper[4818]: I1203 06:46:12.031251 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.031234108 podStartE2EDuration="7.031234108s" podCreationTimestamp="2025-12-03 06:46:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:46:12.027513226 +0000 UTC m=+1129.719122008" watchObservedRunningTime="2025-12-03 06:46:12.031234108 +0000 UTC m=+1129.722842860" Dec 03 06:46:12 crc kubenswrapper[4818]: I1203 06:46:12.977044 4818 generic.go:334] "Generic (PLEG): container finished" podID="a2af4147-3de6-499c-a6f3-133b3e16d673" containerID="fb3177c3e1157b4647a93a74fcb0e22835768f704f8b46e33abb77495a11ddc3" exitCode=0 Dec 03 06:46:12 crc kubenswrapper[4818]: I1203 06:46:12.977330 4818 generic.go:334] "Generic (PLEG): container finished" podID="a2af4147-3de6-499c-a6f3-133b3e16d673" containerID="0cf9e04ee236734f6aa97650dbff35720342c9a3a05bf4c892c2cd049aca1c32" exitCode=143 Dec 03 06:46:12 crc kubenswrapper[4818]: I1203 06:46:12.977224 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2af4147-3de6-499c-a6f3-133b3e16d673","Type":"ContainerDied","Data":"fb3177c3e1157b4647a93a74fcb0e22835768f704f8b46e33abb77495a11ddc3"} Dec 03 06:46:12 crc kubenswrapper[4818]: I1203 06:46:12.977408 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2af4147-3de6-499c-a6f3-133b3e16d673","Type":"ContainerDied","Data":"0cf9e04ee236734f6aa97650dbff35720342c9a3a05bf4c892c2cd049aca1c32"} Dec 03 06:46:12 crc kubenswrapper[4818]: I1203 06:46:12.980574 4818 generic.go:334] "Generic (PLEG): container finished" podID="c50d74e9-8f1f-43f7-ad45-3a9129974cee" containerID="3d9905e829fb27f0002b396cc9c3370b19054e14f905ca2626785b32e804fc2a" exitCode=0 Dec 03 06:46:12 crc kubenswrapper[4818]: I1203 06:46:12.980605 4818 generic.go:334] "Generic (PLEG): container finished" podID="c50d74e9-8f1f-43f7-ad45-3a9129974cee" containerID="98610b9841b1e3b474cfcdf1434dbefe1d8b882c511d2a90727870f68594b9c5" exitCode=143 Dec 03 06:46:12 crc kubenswrapper[4818]: I1203 06:46:12.980624 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c50d74e9-8f1f-43f7-ad45-3a9129974cee","Type":"ContainerDied","Data":"3d9905e829fb27f0002b396cc9c3370b19054e14f905ca2626785b32e804fc2a"} Dec 03 06:46:12 crc kubenswrapper[4818]: I1203 06:46:12.980658 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c50d74e9-8f1f-43f7-ad45-3a9129974cee","Type":"ContainerDied","Data":"98610b9841b1e3b474cfcdf1434dbefe1d8b882c511d2a90727870f68594b9c5"} Dec 03 06:46:13 crc kubenswrapper[4818]: I1203 06:46:13.301950 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:46:13 crc kubenswrapper[4818]: I1203 06:46:13.302552 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:46:13 crc kubenswrapper[4818]: I1203 06:46:13.763980 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:13 crc kubenswrapper[4818]: I1203 06:46:13.934067 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-scripts\") pod \"42277c61-53cd-4617-b5bd-ba45863243d2\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " Dec 03 06:46:13 crc kubenswrapper[4818]: I1203 06:46:13.934126 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-fernet-keys\") pod \"42277c61-53cd-4617-b5bd-ba45863243d2\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " Dec 03 06:46:13 crc kubenswrapper[4818]: I1203 06:46:13.934332 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-credential-keys\") pod \"42277c61-53cd-4617-b5bd-ba45863243d2\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " Dec 03 06:46:13 crc kubenswrapper[4818]: I1203 06:46:13.934405 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-config-data\") pod \"42277c61-53cd-4617-b5bd-ba45863243d2\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " Dec 03 06:46:13 crc kubenswrapper[4818]: I1203 06:46:13.934561 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-combined-ca-bundle\") pod \"42277c61-53cd-4617-b5bd-ba45863243d2\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " Dec 03 06:46:13 crc kubenswrapper[4818]: I1203 06:46:13.934602 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pktv2\" (UniqueName: \"kubernetes.io/projected/42277c61-53cd-4617-b5bd-ba45863243d2-kube-api-access-pktv2\") pod \"42277c61-53cd-4617-b5bd-ba45863243d2\" (UID: \"42277c61-53cd-4617-b5bd-ba45863243d2\") " Dec 03 06:46:13 crc kubenswrapper[4818]: I1203 06:46:13.943913 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-scripts" (OuterVolumeSpecName: "scripts") pod "42277c61-53cd-4617-b5bd-ba45863243d2" (UID: "42277c61-53cd-4617-b5bd-ba45863243d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:13 crc kubenswrapper[4818]: I1203 06:46:13.943946 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "42277c61-53cd-4617-b5bd-ba45863243d2" (UID: "42277c61-53cd-4617-b5bd-ba45863243d2"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:13 crc kubenswrapper[4818]: I1203 06:46:13.944032 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "42277c61-53cd-4617-b5bd-ba45863243d2" (UID: "42277c61-53cd-4617-b5bd-ba45863243d2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:13 crc kubenswrapper[4818]: I1203 06:46:13.988674 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42277c61-53cd-4617-b5bd-ba45863243d2-kube-api-access-pktv2" (OuterVolumeSpecName: "kube-api-access-pktv2") pod "42277c61-53cd-4617-b5bd-ba45863243d2" (UID: "42277c61-53cd-4617-b5bd-ba45863243d2"). InnerVolumeSpecName "kube-api-access-pktv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.029123 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-config-data" (OuterVolumeSpecName: "config-data") pod "42277c61-53cd-4617-b5bd-ba45863243d2" (UID: "42277c61-53cd-4617-b5bd-ba45863243d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.038959 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.038998 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pktv2\" (UniqueName: \"kubernetes.io/projected/42277c61-53cd-4617-b5bd-ba45863243d2-kube-api-access-pktv2\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.039010 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.039020 4818 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.039034 4818 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.088020 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42277c61-53cd-4617-b5bd-ba45863243d2" (UID: "42277c61-53cd-4617-b5bd-ba45863243d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.123792 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4jjs9" event={"ID":"42277c61-53cd-4617-b5bd-ba45863243d2","Type":"ContainerDied","Data":"c4dbb7ed6dec28ce96cc06f3ce4b5897d894ddbf58b71f76ad3d2c99fe295f1c"} Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.123860 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4dbb7ed6dec28ce96cc06f3ce4b5897d894ddbf58b71f76ad3d2c99fe295f1c" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.123942 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4jjs9" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.141273 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42277c61-53cd-4617-b5bd-ba45863243d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.188512 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-4jjs9"] Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.222337 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-4jjs9"] Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.242962 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-v8jgr"] Dec 03 06:46:14 crc kubenswrapper[4818]: E1203 06:46:14.243737 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bfb1795-6ea1-4d0d-bcd1-77937bf698ea" containerName="init" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.243751 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bfb1795-6ea1-4d0d-bcd1-77937bf698ea" containerName="init" Dec 03 06:46:14 crc kubenswrapper[4818]: E1203 06:46:14.243763 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42277c61-53cd-4617-b5bd-ba45863243d2" containerName="keystone-bootstrap" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.243770 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="42277c61-53cd-4617-b5bd-ba45863243d2" containerName="keystone-bootstrap" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.243963 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="42277c61-53cd-4617-b5bd-ba45863243d2" containerName="keystone-bootstrap" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.243987 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bfb1795-6ea1-4d0d-bcd1-77937bf698ea" containerName="init" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.244587 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.247316 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.248242 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.248477 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.252797 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ggd6v" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.252977 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.285826 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v8jgr"] Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.344175 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-combined-ca-bundle\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.344258 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-config-data\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.344282 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srhgb\" (UniqueName: \"kubernetes.io/projected/56e81897-8229-4c21-9beb-a7b0941edf03-kube-api-access-srhgb\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.344345 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-credential-keys\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.344403 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-fernet-keys\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.344452 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-scripts\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.446096 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-credential-keys\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.446211 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-fernet-keys\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.446281 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-scripts\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.446316 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-combined-ca-bundle\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.446365 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-config-data\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.446385 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srhgb\" (UniqueName: \"kubernetes.io/projected/56e81897-8229-4c21-9beb-a7b0941edf03-kube-api-access-srhgb\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.452547 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-config-data\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.452749 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-fernet-keys\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.453281 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-scripts\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.453518 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-combined-ca-bundle\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.454496 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-credential-keys\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.468509 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srhgb\" (UniqueName: \"kubernetes.io/projected/56e81897-8229-4c21-9beb-a7b0941edf03-kube-api-access-srhgb\") pod \"keystone-bootstrap-v8jgr\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.571331 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:14 crc kubenswrapper[4818]: I1203 06:46:14.749961 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42277c61-53cd-4617-b5bd-ba45863243d2" path="/var/lib/kubelet/pods/42277c61-53cd-4617-b5bd-ba45863243d2/volumes" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.325949 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7fbcd7d5fc-6stq9"] Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.356922 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-574dc6486d-nvkzv"] Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.359756 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.371654 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.394148 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-574dc6486d-nvkzv"] Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.454354 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-797779bc9c-df4k9"] Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.465027 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68pvc\" (UniqueName: \"kubernetes.io/projected/bd891b60-48cc-49f1-bb61-9e58c712816e-kube-api-access-68pvc\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.465308 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-combined-ca-bundle\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.465375 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd891b60-48cc-49f1-bb61-9e58c712816e-logs\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.465423 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-horizon-tls-certs\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.465459 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-horizon-secret-key\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.465526 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd891b60-48cc-49f1-bb61-9e58c712816e-scripts\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.465550 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd891b60-48cc-49f1-bb61-9e58c712816e-config-data\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.485304 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-68b68c8c84-qt25s"] Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.487135 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.498051 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68b68c8c84-qt25s"] Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.567067 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68pvc\" (UniqueName: \"kubernetes.io/projected/bd891b60-48cc-49f1-bb61-9e58c712816e-kube-api-access-68pvc\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.567121 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23c89695-46b0-49d8-ab09-eb002d3cb94e-logs\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.567151 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/23c89695-46b0-49d8-ab09-eb002d3cb94e-horizon-secret-key\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.567192 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-combined-ca-bundle\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.567238 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd891b60-48cc-49f1-bb61-9e58c712816e-logs\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.567263 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/23c89695-46b0-49d8-ab09-eb002d3cb94e-horizon-tls-certs\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.567294 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-horizon-tls-certs\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.567323 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-horizon-secret-key\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.567340 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23c89695-46b0-49d8-ab09-eb002d3cb94e-scripts\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.567357 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwnwf\" (UniqueName: \"kubernetes.io/projected/23c89695-46b0-49d8-ab09-eb002d3cb94e-kube-api-access-dwnwf\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.567385 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/23c89695-46b0-49d8-ab09-eb002d3cb94e-config-data\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.567403 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23c89695-46b0-49d8-ab09-eb002d3cb94e-combined-ca-bundle\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.567432 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd891b60-48cc-49f1-bb61-9e58c712816e-scripts\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.567448 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd891b60-48cc-49f1-bb61-9e58c712816e-config-data\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.568437 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd891b60-48cc-49f1-bb61-9e58c712816e-logs\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.568510 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd891b60-48cc-49f1-bb61-9e58c712816e-scripts\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.568945 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd891b60-48cc-49f1-bb61-9e58c712816e-config-data\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.578386 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-combined-ca-bundle\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.580980 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-horizon-secret-key\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.596734 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-horizon-tls-certs\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.602739 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68pvc\" (UniqueName: \"kubernetes.io/projected/bd891b60-48cc-49f1-bb61-9e58c712816e-kube-api-access-68pvc\") pod \"horizon-574dc6486d-nvkzv\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.668791 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/23c89695-46b0-49d8-ab09-eb002d3cb94e-horizon-secret-key\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.669039 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/23c89695-46b0-49d8-ab09-eb002d3cb94e-horizon-tls-certs\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.669109 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23c89695-46b0-49d8-ab09-eb002d3cb94e-scripts\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.669149 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwnwf\" (UniqueName: \"kubernetes.io/projected/23c89695-46b0-49d8-ab09-eb002d3cb94e-kube-api-access-dwnwf\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.669190 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/23c89695-46b0-49d8-ab09-eb002d3cb94e-config-data\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.669213 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23c89695-46b0-49d8-ab09-eb002d3cb94e-combined-ca-bundle\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.669287 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23c89695-46b0-49d8-ab09-eb002d3cb94e-logs\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.669753 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23c89695-46b0-49d8-ab09-eb002d3cb94e-logs\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.670287 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23c89695-46b0-49d8-ab09-eb002d3cb94e-scripts\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.672034 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/23c89695-46b0-49d8-ab09-eb002d3cb94e-horizon-secret-key\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.673122 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/23c89695-46b0-49d8-ab09-eb002d3cb94e-horizon-tls-certs\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.678481 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/23c89695-46b0-49d8-ab09-eb002d3cb94e-config-data\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.679192 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23c89695-46b0-49d8-ab09-eb002d3cb94e-combined-ca-bundle\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.683601 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.689573 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwnwf\" (UniqueName: \"kubernetes.io/projected/23c89695-46b0-49d8-ab09-eb002d3cb94e-kube-api-access-dwnwf\") pod \"horizon-68b68c8c84-qt25s\" (UID: \"23c89695-46b0-49d8-ab09-eb002d3cb94e\") " pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:15 crc kubenswrapper[4818]: I1203 06:46:15.815496 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:16 crc kubenswrapper[4818]: I1203 06:46:16.302028 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:16 crc kubenswrapper[4818]: I1203 06:46:16.370439 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sfd82"] Dec 03 06:46:16 crc kubenswrapper[4818]: I1203 06:46:16.370720 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" podUID="4803b078-4593-4b55-8a42-f4c3f4b0e21f" containerName="dnsmasq-dns" containerID="cri-o://7e1e0db439eb910a33d0397c6942599f54f43b4b79841c87971413ca609fdb74" gracePeriod=10 Dec 03 06:46:17 crc kubenswrapper[4818]: I1203 06:46:17.158351 4818 generic.go:334] "Generic (PLEG): container finished" podID="4803b078-4593-4b55-8a42-f4c3f4b0e21f" containerID="7e1e0db439eb910a33d0397c6942599f54f43b4b79841c87971413ca609fdb74" exitCode=0 Dec 03 06:46:17 crc kubenswrapper[4818]: I1203 06:46:17.158539 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" event={"ID":"4803b078-4593-4b55-8a42-f4c3f4b0e21f","Type":"ContainerDied","Data":"7e1e0db439eb910a33d0397c6942599f54f43b4b79841c87971413ca609fdb74"} Dec 03 06:46:19 crc kubenswrapper[4818]: I1203 06:46:19.084811 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" podUID="4803b078-4593-4b55-8a42-f4c3f4b0e21f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.131:5353: connect: connection refused" Dec 03 06:46:19 crc kubenswrapper[4818]: E1203 06:46:19.562073 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e751691_c660_4527_a9a6_61505f83c6ff.slice/crio-303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0.scope\": RecentStats: unable to find data in memory cache]" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.110128 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.274867 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a2af4147-3de6-499c-a6f3-133b3e16d673","Type":"ContainerDied","Data":"984be78d81560ce343a6cc21d1c7c233b96d9c10c2be88a9ae778816ab387230"} Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.274932 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.274942 4818 scope.go:117] "RemoveContainer" containerID="fb3177c3e1157b4647a93a74fcb0e22835768f704f8b46e33abb77495a11ddc3" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.278316 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-combined-ca-bundle\") pod \"a2af4147-3de6-499c-a6f3-133b3e16d673\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.278392 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"a2af4147-3de6-499c-a6f3-133b3e16d673\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.278423 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2af4147-3de6-499c-a6f3-133b3e16d673-httpd-run\") pod \"a2af4147-3de6-499c-a6f3-133b3e16d673\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.278504 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2af4147-3de6-499c-a6f3-133b3e16d673-logs\") pod \"a2af4147-3de6-499c-a6f3-133b3e16d673\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.278563 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-scripts\") pod \"a2af4147-3de6-499c-a6f3-133b3e16d673\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.278584 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-config-data\") pod \"a2af4147-3de6-499c-a6f3-133b3e16d673\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.278700 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9kh5\" (UniqueName: \"kubernetes.io/projected/a2af4147-3de6-499c-a6f3-133b3e16d673-kube-api-access-d9kh5\") pod \"a2af4147-3de6-499c-a6f3-133b3e16d673\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.278722 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-internal-tls-certs\") pod \"a2af4147-3de6-499c-a6f3-133b3e16d673\" (UID: \"a2af4147-3de6-499c-a6f3-133b3e16d673\") " Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.279828 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2af4147-3de6-499c-a6f3-133b3e16d673-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a2af4147-3de6-499c-a6f3-133b3e16d673" (UID: "a2af4147-3de6-499c-a6f3-133b3e16d673"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.279984 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2af4147-3de6-499c-a6f3-133b3e16d673-logs" (OuterVolumeSpecName: "logs") pod "a2af4147-3de6-499c-a6f3-133b3e16d673" (UID: "a2af4147-3de6-499c-a6f3-133b3e16d673"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.284385 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2af4147-3de6-499c-a6f3-133b3e16d673-kube-api-access-d9kh5" (OuterVolumeSpecName: "kube-api-access-d9kh5") pod "a2af4147-3de6-499c-a6f3-133b3e16d673" (UID: "a2af4147-3de6-499c-a6f3-133b3e16d673"). InnerVolumeSpecName "kube-api-access-d9kh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.284475 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "a2af4147-3de6-499c-a6f3-133b3e16d673" (UID: "a2af4147-3de6-499c-a6f3-133b3e16d673"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.299492 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-scripts" (OuterVolumeSpecName: "scripts") pod "a2af4147-3de6-499c-a6f3-133b3e16d673" (UID: "a2af4147-3de6-499c-a6f3-133b3e16d673"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.316633 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a2af4147-3de6-499c-a6f3-133b3e16d673" (UID: "a2af4147-3de6-499c-a6f3-133b3e16d673"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.330043 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a2af4147-3de6-499c-a6f3-133b3e16d673" (UID: "a2af4147-3de6-499c-a6f3-133b3e16d673"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.333228 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-config-data" (OuterVolumeSpecName: "config-data") pod "a2af4147-3de6-499c-a6f3-133b3e16d673" (UID: "a2af4147-3de6-499c-a6f3-133b3e16d673"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.380467 4818 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.380507 4818 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2af4147-3de6-499c-a6f3-133b3e16d673-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.380520 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2af4147-3de6-499c-a6f3-133b3e16d673-logs\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.380531 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.380543 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.380556 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9kh5\" (UniqueName: \"kubernetes.io/projected/a2af4147-3de6-499c-a6f3-133b3e16d673-kube-api-access-d9kh5\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.380568 4818 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.380579 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2af4147-3de6-499c-a6f3-133b3e16d673-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.399949 4818 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.482290 4818 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.619051 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.626289 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.645281 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 06:46:20 crc kubenswrapper[4818]: E1203 06:46:20.645939 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2af4147-3de6-499c-a6f3-133b3e16d673" containerName="glance-httpd" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.646006 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2af4147-3de6-499c-a6f3-133b3e16d673" containerName="glance-httpd" Dec 03 06:46:20 crc kubenswrapper[4818]: E1203 06:46:20.646074 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2af4147-3de6-499c-a6f3-133b3e16d673" containerName="glance-log" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.646119 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2af4147-3de6-499c-a6f3-133b3e16d673" containerName="glance-log" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.646321 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2af4147-3de6-499c-a6f3-133b3e16d673" containerName="glance-log" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.646387 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2af4147-3de6-499c-a6f3-133b3e16d673" containerName="glance-httpd" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.649559 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.658498 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.658807 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.663127 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.751101 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2af4147-3de6-499c-a6f3-133b3e16d673" path="/var/lib/kubelet/pods/a2af4147-3de6-499c-a6f3-133b3e16d673/volumes" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.793486 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.793532 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g4mg\" (UniqueName: \"kubernetes.io/projected/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-kube-api-access-6g4mg\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.793563 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.793590 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.793623 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.793774 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.793804 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.793852 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-logs\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.896284 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.895714 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.896637 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.897052 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-logs\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.897447 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-logs\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.897632 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.898214 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g4mg\" (UniqueName: \"kubernetes.io/projected/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-kube-api-access-6g4mg\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.898624 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.898682 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.898724 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.899078 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.905871 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.905888 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.906605 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.916527 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g4mg\" (UniqueName: \"kubernetes.io/projected/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-kube-api-access-6g4mg\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.918524 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.939196 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:46:20 crc kubenswrapper[4818]: I1203 06:46:20.987047 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 06:46:22 crc kubenswrapper[4818]: E1203 06:46:22.479529 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 03 06:46:22 crc kubenswrapper[4818]: E1203 06:46:22.479924 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n549h64dhf5hc9hf6h65dh65ch5f4hcbh586hd8h84h5b9hbch647hdh588h65h7ch56h85h644hcfh67fhf7h55dh5cbh548h5d8hf4hc9h56fq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5d2sg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(575fe180-2a23-48e9-b5cf-b90e83528e5e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:46:22 crc kubenswrapper[4818]: E1203 06:46:22.489100 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 03 06:46:22 crc kubenswrapper[4818]: E1203 06:46:22.489323 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n65dhf4h695h565h67ch676h675h575h5dbh5cch58h66fh5c4hcch59fh59bhd6h647h566h8h5bbhc9h54bh64chb5h95h96h67fhb4h8dh8bh557q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xgpl2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5649b85449-wr6w6_openstack(38ad07da-c33a-41ff-96c4-dce7dfa8d3f9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:46:22 crc kubenswrapper[4818]: E1203 06:46:22.491971 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5649b85449-wr6w6" podUID="38ad07da-c33a-41ff-96c4-dce7dfa8d3f9" Dec 03 06:46:29 crc kubenswrapper[4818]: I1203 06:46:29.084674 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" podUID="4803b078-4593-4b55-8a42-f4c3f4b0e21f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.131:5353: i/o timeout" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.079927 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.085914 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.090137 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.186682 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-combined-ca-bundle\") pod \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.186728 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-config-data\") pod \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.186750 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-ovsdbserver-nb\") pod \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.186777 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-public-tls-certs\") pod \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.186800 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-logs\") pod \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.186862 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.186904 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdkrp\" (UniqueName: \"kubernetes.io/projected/4803b078-4593-4b55-8a42-f4c3f4b0e21f-kube-api-access-gdkrp\") pod \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.186961 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-ovsdbserver-sb\") pod \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.186997 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79bkg\" (UniqueName: \"kubernetes.io/projected/c50d74e9-8f1f-43f7-ad45-3a9129974cee-kube-api-access-79bkg\") pod \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.187396 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-dns-swift-storage-0\") pod \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.187440 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgpl2\" (UniqueName: \"kubernetes.io/projected/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-kube-api-access-xgpl2\") pod \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.187510 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c50d74e9-8f1f-43f7-ad45-3a9129974cee-httpd-run\") pod \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.187494 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-logs" (OuterVolumeSpecName: "logs") pod "38ad07da-c33a-41ff-96c4-dce7dfa8d3f9" (UID: "38ad07da-c33a-41ff-96c4-dce7dfa8d3f9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.187570 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-horizon-secret-key\") pod \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.187609 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-scripts\") pod \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.187695 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-scripts\") pod \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.187730 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-dns-svc\") pod \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.187751 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c50d74e9-8f1f-43f7-ad45-3a9129974cee-logs\") pod \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\" (UID: \"c50d74e9-8f1f-43f7-ad45-3a9129974cee\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.187806 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-config-data\") pod \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\" (UID: \"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.187873 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-config\") pod \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\" (UID: \"4803b078-4593-4b55-8a42-f4c3f4b0e21f\") " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.188420 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-logs\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.188582 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-scripts" (OuterVolumeSpecName: "scripts") pod "38ad07da-c33a-41ff-96c4-dce7dfa8d3f9" (UID: "38ad07da-c33a-41ff-96c4-dce7dfa8d3f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.188878 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c50d74e9-8f1f-43f7-ad45-3a9129974cee-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c50d74e9-8f1f-43f7-ad45-3a9129974cee" (UID: "c50d74e9-8f1f-43f7-ad45-3a9129974cee"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.192727 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "38ad07da-c33a-41ff-96c4-dce7dfa8d3f9" (UID: "38ad07da-c33a-41ff-96c4-dce7dfa8d3f9"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.193033 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c50d74e9-8f1f-43f7-ad45-3a9129974cee-logs" (OuterVolumeSpecName: "logs") pod "c50d74e9-8f1f-43f7-ad45-3a9129974cee" (UID: "c50d74e9-8f1f-43f7-ad45-3a9129974cee"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.193143 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4803b078-4593-4b55-8a42-f4c3f4b0e21f-kube-api-access-gdkrp" (OuterVolumeSpecName: "kube-api-access-gdkrp") pod "4803b078-4593-4b55-8a42-f4c3f4b0e21f" (UID: "4803b078-4593-4b55-8a42-f4c3f4b0e21f"). InnerVolumeSpecName "kube-api-access-gdkrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.193478 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "c50d74e9-8f1f-43f7-ad45-3a9129974cee" (UID: "c50d74e9-8f1f-43f7-ad45-3a9129974cee"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.193743 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-config-data" (OuterVolumeSpecName: "config-data") pod "38ad07da-c33a-41ff-96c4-dce7dfa8d3f9" (UID: "38ad07da-c33a-41ff-96c4-dce7dfa8d3f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.196230 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c50d74e9-8f1f-43f7-ad45-3a9129974cee-kube-api-access-79bkg" (OuterVolumeSpecName: "kube-api-access-79bkg") pod "c50d74e9-8f1f-43f7-ad45-3a9129974cee" (UID: "c50d74e9-8f1f-43f7-ad45-3a9129974cee"). InnerVolumeSpecName "kube-api-access-79bkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.206760 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-scripts" (OuterVolumeSpecName: "scripts") pod "c50d74e9-8f1f-43f7-ad45-3a9129974cee" (UID: "c50d74e9-8f1f-43f7-ad45-3a9129974cee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.209961 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-kube-api-access-xgpl2" (OuterVolumeSpecName: "kube-api-access-xgpl2") pod "38ad07da-c33a-41ff-96c4-dce7dfa8d3f9" (UID: "38ad07da-c33a-41ff-96c4-dce7dfa8d3f9"). InnerVolumeSpecName "kube-api-access-xgpl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.233952 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c50d74e9-8f1f-43f7-ad45-3a9129974cee" (UID: "c50d74e9-8f1f-43f7-ad45-3a9129974cee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.259749 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4803b078-4593-4b55-8a42-f4c3f4b0e21f" (UID: "4803b078-4593-4b55-8a42-f4c3f4b0e21f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.267567 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4803b078-4593-4b55-8a42-f4c3f4b0e21f" (UID: "4803b078-4593-4b55-8a42-f4c3f4b0e21f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.276644 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4803b078-4593-4b55-8a42-f4c3f4b0e21f" (UID: "4803b078-4593-4b55-8a42-f4c3f4b0e21f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.278492 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-config-data" (OuterVolumeSpecName: "config-data") pod "c50d74e9-8f1f-43f7-ad45-3a9129974cee" (UID: "c50d74e9-8f1f-43f7-ad45-3a9129974cee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.282211 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c50d74e9-8f1f-43f7-ad45-3a9129974cee" (UID: "c50d74e9-8f1f-43f7-ad45-3a9129974cee"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.290004 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79bkg\" (UniqueName: \"kubernetes.io/projected/c50d74e9-8f1f-43f7-ad45-3a9129974cee-kube-api-access-79bkg\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.290048 4818 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.290062 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgpl2\" (UniqueName: \"kubernetes.io/projected/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-kube-api-access-xgpl2\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.290077 4818 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c50d74e9-8f1f-43f7-ad45-3a9129974cee-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.290088 4818 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.290099 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.290110 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.290120 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.290130 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c50d74e9-8f1f-43f7-ad45-3a9129974cee-logs\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.290140 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.290152 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.290163 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.290174 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.290185 4818 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c50d74e9-8f1f-43f7-ad45-3a9129974cee-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.290223 4818 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.290236 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdkrp\" (UniqueName: \"kubernetes.io/projected/4803b078-4593-4b55-8a42-f4c3f4b0e21f-kube-api-access-gdkrp\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.290384 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4803b078-4593-4b55-8a42-f4c3f4b0e21f" (UID: "4803b078-4593-4b55-8a42-f4c3f4b0e21f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.307591 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-config" (OuterVolumeSpecName: "config") pod "4803b078-4593-4b55-8a42-f4c3f4b0e21f" (UID: "4803b078-4593-4b55-8a42-f4c3f4b0e21f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.313059 4818 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.391605 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.391653 4818 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.391664 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4803b078-4593-4b55-8a42-f4c3f4b0e21f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.407503 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5649b85449-wr6w6" event={"ID":"38ad07da-c33a-41ff-96c4-dce7dfa8d3f9","Type":"ContainerDied","Data":"850ead00d645ec72cc20fd3abd75672540a69c0fe3b9f172e911ac43851dac62"} Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.407522 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5649b85449-wr6w6" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.409722 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" event={"ID":"4803b078-4593-4b55-8a42-f4c3f4b0e21f","Type":"ContainerDied","Data":"99a4375a31a3b06f220595540cb9121692a2490d26e646ddea9c1148f74e88f1"} Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.409777 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.412175 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c50d74e9-8f1f-43f7-ad45-3a9129974cee","Type":"ContainerDied","Data":"c1b13df7785d17fa5a6657f602efc94d7b5f9a6edc5bd4c712539587e65a3395"} Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.412224 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.464511 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sfd82"] Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.489088 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sfd82"] Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.501666 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.512213 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.523278 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 06:46:33 crc kubenswrapper[4818]: E1203 06:46:33.524068 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c50d74e9-8f1f-43f7-ad45-3a9129974cee" containerName="glance-httpd" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.524095 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c50d74e9-8f1f-43f7-ad45-3a9129974cee" containerName="glance-httpd" Dec 03 06:46:33 crc kubenswrapper[4818]: E1203 06:46:33.524114 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c50d74e9-8f1f-43f7-ad45-3a9129974cee" containerName="glance-log" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.524123 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c50d74e9-8f1f-43f7-ad45-3a9129974cee" containerName="glance-log" Dec 03 06:46:33 crc kubenswrapper[4818]: E1203 06:46:33.524137 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4803b078-4593-4b55-8a42-f4c3f4b0e21f" containerName="dnsmasq-dns" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.524145 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4803b078-4593-4b55-8a42-f4c3f4b0e21f" containerName="dnsmasq-dns" Dec 03 06:46:33 crc kubenswrapper[4818]: E1203 06:46:33.524159 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4803b078-4593-4b55-8a42-f4c3f4b0e21f" containerName="init" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.524167 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4803b078-4593-4b55-8a42-f4c3f4b0e21f" containerName="init" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.524391 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="c50d74e9-8f1f-43f7-ad45-3a9129974cee" containerName="glance-httpd" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.524413 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4803b078-4593-4b55-8a42-f4c3f4b0e21f" containerName="dnsmasq-dns" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.524432 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="c50d74e9-8f1f-43f7-ad45-3a9129974cee" containerName="glance-log" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.525651 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.527798 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.528047 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.550636 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5649b85449-wr6w6"] Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.558845 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5649b85449-wr6w6"] Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.566135 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.696568 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.696635 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.696688 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-config-data\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.697151 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58r5k\" (UniqueName: \"kubernetes.io/projected/34a77ce6-1705-45d6-af7f-8100e13a6d3b-kube-api-access-58r5k\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.697180 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-scripts\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.697261 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/34a77ce6-1705-45d6-af7f-8100e13a6d3b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.697467 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.697559 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34a77ce6-1705-45d6-af7f-8100e13a6d3b-logs\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: E1203 06:46:33.768191 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 03 06:46:33 crc kubenswrapper[4818]: E1203 06:46:33.768728 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vwqv8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-7kbvz_openstack(982110da-ac09-4df8-9729-77284ed7539b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:46:33 crc kubenswrapper[4818]: E1203 06:46:33.770399 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-7kbvz" podUID="982110da-ac09-4df8-9729-77284ed7539b" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.798925 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58r5k\" (UniqueName: \"kubernetes.io/projected/34a77ce6-1705-45d6-af7f-8100e13a6d3b-kube-api-access-58r5k\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.798967 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-scripts\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.798985 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/34a77ce6-1705-45d6-af7f-8100e13a6d3b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.799045 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.799079 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34a77ce6-1705-45d6-af7f-8100e13a6d3b-logs\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.799094 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.799123 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.799137 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-config-data\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.799540 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34a77ce6-1705-45d6-af7f-8100e13a6d3b-logs\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.799658 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.799803 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/34a77ce6-1705-45d6-af7f-8100e13a6d3b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.805558 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.806239 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-config-data\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.807089 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.813241 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-scripts\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.820627 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58r5k\" (UniqueName: \"kubernetes.io/projected/34a77ce6-1705-45d6-af7f-8100e13a6d3b-kube-api-access-58r5k\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.826342 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " pod="openstack/glance-default-external-api-0" Dec 03 06:46:33 crc kubenswrapper[4818]: I1203 06:46:33.845871 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 06:46:34 crc kubenswrapper[4818]: I1203 06:46:34.086096 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-sfd82" podUID="4803b078-4593-4b55-8a42-f4c3f4b0e21f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.131:5353: i/o timeout" Dec 03 06:46:34 crc kubenswrapper[4818]: I1203 06:46:34.422945 4818 generic.go:334] "Generic (PLEG): container finished" podID="cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7" containerID="a1d52d0e41b793fc88a51dc145961cc2a6b3b6f46b3ca4f38a0e30404c4e6cdf" exitCode=0 Dec 03 06:46:34 crc kubenswrapper[4818]: I1203 06:46:34.423024 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-ffzld" event={"ID":"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7","Type":"ContainerDied","Data":"a1d52d0e41b793fc88a51dc145961cc2a6b3b6f46b3ca4f38a0e30404c4e6cdf"} Dec 03 06:46:34 crc kubenswrapper[4818]: E1203 06:46:34.424403 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-7kbvz" podUID="982110da-ac09-4df8-9729-77284ed7539b" Dec 03 06:46:34 crc kubenswrapper[4818]: I1203 06:46:34.748633 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38ad07da-c33a-41ff-96c4-dce7dfa8d3f9" path="/var/lib/kubelet/pods/38ad07da-c33a-41ff-96c4-dce7dfa8d3f9/volumes" Dec 03 06:46:34 crc kubenswrapper[4818]: I1203 06:46:34.749107 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4803b078-4593-4b55-8a42-f4c3f4b0e21f" path="/var/lib/kubelet/pods/4803b078-4593-4b55-8a42-f4c3f4b0e21f/volumes" Dec 03 06:46:34 crc kubenswrapper[4818]: I1203 06:46:34.750116 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c50d74e9-8f1f-43f7-ad45-3a9129974cee" path="/var/lib/kubelet/pods/c50d74e9-8f1f-43f7-ad45-3a9129974cee/volumes" Dec 03 06:46:34 crc kubenswrapper[4818]: I1203 06:46:34.782968 4818 scope.go:117] "RemoveContainer" containerID="0cf9e04ee236734f6aa97650dbff35720342c9a3a05bf4c892c2cd049aca1c32" Dec 03 06:46:34 crc kubenswrapper[4818]: E1203 06:46:34.800113 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 03 06:46:34 crc kubenswrapper[4818]: E1203 06:46:34.800277 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xhh5p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-978k5_openstack(20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 06:46:34 crc kubenswrapper[4818]: E1203 06:46:34.801436 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-978k5" podUID="20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3" Dec 03 06:46:35 crc kubenswrapper[4818]: I1203 06:46:35.184672 4818 scope.go:117] "RemoveContainer" containerID="7e1e0db439eb910a33d0397c6942599f54f43b4b79841c87971413ca609fdb74" Dec 03 06:46:35 crc kubenswrapper[4818]: I1203 06:46:35.240105 4818 scope.go:117] "RemoveContainer" containerID="0e0da409a630e9d2e62668e45773b61af258dc6e15a86d708d0cad1f31f007ac" Dec 03 06:46:35 crc kubenswrapper[4818]: I1203 06:46:35.321832 4818 scope.go:117] "RemoveContainer" containerID="3d9905e829fb27f0002b396cc9c3370b19054e14f905ca2626785b32e804fc2a" Dec 03 06:46:35 crc kubenswrapper[4818]: I1203 06:46:35.361885 4818 scope.go:117] "RemoveContainer" containerID="98610b9841b1e3b474cfcdf1434dbefe1d8b882c511d2a90727870f68594b9c5" Dec 03 06:46:35 crc kubenswrapper[4818]: I1203 06:46:35.433600 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-797779bc9c-df4k9" event={"ID":"2f18580e-5450-4edb-b266-734372950182","Type":"ContainerStarted","Data":"9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc"} Dec 03 06:46:35 crc kubenswrapper[4818]: I1203 06:46:35.435921 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lqzrp" event={"ID":"c23baebd-4d73-4972-b9eb-5c573e859977","Type":"ContainerStarted","Data":"a20b9ba01523bf083f62485b116ef7085bb29d10e0791cfce5c65d1bbc2107f9"} Dec 03 06:46:35 crc kubenswrapper[4818]: E1203 06:46:35.440567 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-978k5" podUID="20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3" Dec 03 06:46:35 crc kubenswrapper[4818]: I1203 06:46:35.456231 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-lqzrp" podStartSLOduration=4.860789517 podStartE2EDuration="30.456209805s" podCreationTimestamp="2025-12-03 06:46:05 +0000 UTC" firstStartedPulling="2025-12-03 06:46:07.395253142 +0000 UTC m=+1125.086861894" lastFinishedPulling="2025-12-03 06:46:32.99067342 +0000 UTC m=+1150.682282182" observedRunningTime="2025-12-03 06:46:35.453530569 +0000 UTC m=+1153.145139321" watchObservedRunningTime="2025-12-03 06:46:35.456209805 +0000 UTC m=+1153.147818557" Dec 03 06:46:35 crc kubenswrapper[4818]: I1203 06:46:35.527119 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v8jgr"] Dec 03 06:46:35 crc kubenswrapper[4818]: W1203 06:46:35.545231 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56e81897_8229_4c21_9beb_a7b0941edf03.slice/crio-dd918ad3067671c1127691fb2afa4ae2c4b58a5c4c3cc1040b4cfe018533952e WatchSource:0}: Error finding container dd918ad3067671c1127691fb2afa4ae2c4b58a5c4c3cc1040b4cfe018533952e: Status 404 returned error can't find the container with id dd918ad3067671c1127691fb2afa4ae2c4b58a5c4c3cc1040b4cfe018533952e Dec 03 06:46:35 crc kubenswrapper[4818]: I1203 06:46:35.552100 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 06:46:35 crc kubenswrapper[4818]: I1203 06:46:35.636976 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-574dc6486d-nvkzv"] Dec 03 06:46:35 crc kubenswrapper[4818]: W1203 06:46:35.641432 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd891b60_48cc_49f1_bb61_9e58c712816e.slice/crio-40aec6581deeb30113d6bd25b29b2eb16765b914a7708455f81e07f5041b88c0 WatchSource:0}: Error finding container 40aec6581deeb30113d6bd25b29b2eb16765b914a7708455f81e07f5041b88c0: Status 404 returned error can't find the container with id 40aec6581deeb30113d6bd25b29b2eb16765b914a7708455f81e07f5041b88c0 Dec 03 06:46:35 crc kubenswrapper[4818]: I1203 06:46:35.657647 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68b68c8c84-qt25s"] Dec 03 06:46:35 crc kubenswrapper[4818]: I1203 06:46:35.742712 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 06:46:35 crc kubenswrapper[4818]: W1203 06:46:35.757014 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11dff550_2d3d_4bbc_bbf8_e1bb6849c94f.slice/crio-6ffe09976590646045227ddb478e5fd8d01c9354f24f473bc74dabce43fdd282 WatchSource:0}: Error finding container 6ffe09976590646045227ddb478e5fd8d01c9354f24f473bc74dabce43fdd282: Status 404 returned error can't find the container with id 6ffe09976590646045227ddb478e5fd8d01c9354f24f473bc74dabce43fdd282 Dec 03 06:46:35 crc kubenswrapper[4818]: I1203 06:46:35.850896 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-ffzld" Dec 03 06:46:35 crc kubenswrapper[4818]: I1203 06:46:35.950045 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-config\") pod \"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7\" (UID: \"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7\") " Dec 03 06:46:35 crc kubenswrapper[4818]: I1203 06:46:35.950116 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-combined-ca-bundle\") pod \"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7\" (UID: \"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7\") " Dec 03 06:46:35 crc kubenswrapper[4818]: I1203 06:46:35.950202 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dv774\" (UniqueName: \"kubernetes.io/projected/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-kube-api-access-dv774\") pod \"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7\" (UID: \"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7\") " Dec 03 06:46:35 crc kubenswrapper[4818]: I1203 06:46:35.956196 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-kube-api-access-dv774" (OuterVolumeSpecName: "kube-api-access-dv774") pod "cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7" (UID: "cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7"). InnerVolumeSpecName "kube-api-access-dv774". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.006498 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7" (UID: "cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.037518 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-config" (OuterVolumeSpecName: "config") pod "cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7" (UID: "cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.046354 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.052195 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.052236 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.052253 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dv774\" (UniqueName: \"kubernetes.io/projected/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7-kube-api-access-dv774\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.487207 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"34a77ce6-1705-45d6-af7f-8100e13a6d3b","Type":"ContainerStarted","Data":"7ba1a8fab1700ed6f9676c759f75dc7bf7d9ca4bb8dcf07630cb26aa3bd0c287"} Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.497765 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"575fe180-2a23-48e9-b5cf-b90e83528e5e","Type":"ContainerStarted","Data":"8f699e105a92636cbeab2633821e2e90792849bdc5323a39d6e9a3ee12f8fa9c"} Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.508205 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-ffzld" event={"ID":"cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7","Type":"ContainerDied","Data":"4bfa7ccda7f77d76153ede6ac6a2ccddc8f21655beef1f0b5cae25a27aa84758"} Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.508246 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bfa7ccda7f77d76153ede6ac6a2ccddc8f21655beef1f0b5cae25a27aa84758" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.508315 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-ffzld" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.521926 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v8jgr" event={"ID":"56e81897-8229-4c21-9beb-a7b0941edf03","Type":"ContainerStarted","Data":"19a34de912c8415893becf259c87b4bece98c0984dc27039b91e46ed8aeb25b5"} Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.522248 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v8jgr" event={"ID":"56e81897-8229-4c21-9beb-a7b0941edf03","Type":"ContainerStarted","Data":"dd918ad3067671c1127691fb2afa4ae2c4b58a5c4c3cc1040b4cfe018533952e"} Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.540763 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fbcd7d5fc-6stq9" event={"ID":"9d0cba79-7a13-40ee-ab38-ae093b204a9c","Type":"ContainerStarted","Data":"42d52b447fd707dfb9a71ccd0ff746cf0e8dda0c5ba69b869f28d7f7cd076848"} Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.540807 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fbcd7d5fc-6stq9" event={"ID":"9d0cba79-7a13-40ee-ab38-ae093b204a9c","Type":"ContainerStarted","Data":"703567b992f5fda96a6c66e4c3c5c7af52da72a3847c84b247816dc6adfdcd7a"} Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.541003 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7fbcd7d5fc-6stq9" podUID="9d0cba79-7a13-40ee-ab38-ae093b204a9c" containerName="horizon-log" containerID="cri-o://703567b992f5fda96a6c66e4c3c5c7af52da72a3847c84b247816dc6adfdcd7a" gracePeriod=30 Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.541136 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7fbcd7d5fc-6stq9" podUID="9d0cba79-7a13-40ee-ab38-ae093b204a9c" containerName="horizon" containerID="cri-o://42d52b447fd707dfb9a71ccd0ff746cf0e8dda0c5ba69b869f28d7f7cd076848" gracePeriod=30 Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.546519 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-797779bc9c-df4k9" event={"ID":"2f18580e-5450-4edb-b266-734372950182","Type":"ContainerStarted","Data":"707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115"} Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.546642 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-797779bc9c-df4k9" podUID="2f18580e-5450-4edb-b266-734372950182" containerName="horizon-log" containerID="cri-o://9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc" gracePeriod=30 Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.546728 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-797779bc9c-df4k9" podUID="2f18580e-5450-4edb-b266-734372950182" containerName="horizon" containerID="cri-o://707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115" gracePeriod=30 Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.554035 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-v8jgr" podStartSLOduration=22.554013411 podStartE2EDuration="22.554013411s" podCreationTimestamp="2025-12-03 06:46:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:46:36.537198106 +0000 UTC m=+1154.228806858" watchObservedRunningTime="2025-12-03 06:46:36.554013411 +0000 UTC m=+1154.245622163" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.556222 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574dc6486d-nvkzv" event={"ID":"bd891b60-48cc-49f1-bb61-9e58c712816e","Type":"ContainerStarted","Data":"1ced693b4186c9669b7fde7dbc2e7323ca419d09850500ca4d98afd3fb031afe"} Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.556294 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574dc6486d-nvkzv" event={"ID":"bd891b60-48cc-49f1-bb61-9e58c712816e","Type":"ContainerStarted","Data":"40aec6581deeb30113d6bd25b29b2eb16765b914a7708455f81e07f5041b88c0"} Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.578089 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7fbcd7d5fc-6stq9" podStartSLOduration=3.513280278 podStartE2EDuration="31.578058976s" podCreationTimestamp="2025-12-03 06:46:05 +0000 UTC" firstStartedPulling="2025-12-03 06:46:07.054997498 +0000 UTC m=+1124.746606250" lastFinishedPulling="2025-12-03 06:46:35.119776196 +0000 UTC m=+1152.811384948" observedRunningTime="2025-12-03 06:46:36.560204935 +0000 UTC m=+1154.251813687" watchObservedRunningTime="2025-12-03 06:46:36.578058976 +0000 UTC m=+1154.269667728" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.580915 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f","Type":"ContainerStarted","Data":"6ffe09976590646045227ddb478e5fd8d01c9354f24f473bc74dabce43fdd282"} Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.594227 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68b68c8c84-qt25s" event={"ID":"23c89695-46b0-49d8-ab09-eb002d3cb94e","Type":"ContainerStarted","Data":"dc5d264acdd2b3a88ce87be481183851c0f2f7f54357087e4a98cb253d1779a4"} Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.594278 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68b68c8c84-qt25s" event={"ID":"23c89695-46b0-49d8-ab09-eb002d3cb94e","Type":"ContainerStarted","Data":"5b0eece6f61dd6e2cd41a80becd22568be0c5e6d73cbfef2d5c92fafa6d83a1e"} Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.612968 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-574dc6486d-nvkzv" podStartSLOduration=21.612941369 podStartE2EDuration="21.612941369s" podCreationTimestamp="2025-12-03 06:46:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:46:36.586149576 +0000 UTC m=+1154.277758328" watchObservedRunningTime="2025-12-03 06:46:36.612941369 +0000 UTC m=+1154.304550131" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.698467 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-797779bc9c-df4k9" podStartSLOduration=5.082393722 podStartE2EDuration="29.698441773s" podCreationTimestamp="2025-12-03 06:46:07 +0000 UTC" firstStartedPulling="2025-12-03 06:46:10.167065701 +0000 UTC m=+1127.858674453" lastFinishedPulling="2025-12-03 06:46:34.783113752 +0000 UTC m=+1152.474722504" observedRunningTime="2025-12-03 06:46:36.609452732 +0000 UTC m=+1154.301061504" watchObservedRunningTime="2025-12-03 06:46:36.698441773 +0000 UTC m=+1154.390050525" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.763260 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-68b68c8c84-qt25s" podStartSLOduration=21.763238095 podStartE2EDuration="21.763238095s" podCreationTimestamp="2025-12-03 06:46:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:46:36.754591381 +0000 UTC m=+1154.446200143" watchObservedRunningTime="2025-12-03 06:46:36.763238095 +0000 UTC m=+1154.454846847" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.789338 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-pb59n"] Dec 03 06:46:36 crc kubenswrapper[4818]: E1203 06:46:36.790134 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7" containerName="neutron-db-sync" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.790252 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7" containerName="neutron-db-sync" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.790486 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7" containerName="neutron-db-sync" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.792204 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.833893 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-pb59n"] Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.913513 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk7x9\" (UniqueName: \"kubernetes.io/projected/8cafb812-b94e-45b2-9f13-9a2be0599382-kube-api-access-mk7x9\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.914032 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-dns-svc\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.914268 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.914441 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-config\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.914615 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.914934 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:36 crc kubenswrapper[4818]: I1203 06:46:36.995516 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-69b59f8dd-sckpz"] Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.000840 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.018422 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.019663 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69b59f8dd-sckpz"] Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.018861 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-zctkr" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.018960 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.019703 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.031047 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-config\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.037489 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.032228 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-config\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.040205 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.041000 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.044266 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk7x9\" (UniqueName: \"kubernetes.io/projected/8cafb812-b94e-45b2-9f13-9a2be0599382-kube-api-access-mk7x9\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.044658 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-dns-svc\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.045331 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.043617 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.049565 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.054765 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-dns-svc\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.097226 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk7x9\" (UniqueName: \"kubernetes.io/projected/8cafb812-b94e-45b2-9f13-9a2be0599382-kube-api-access-mk7x9\") pod \"dnsmasq-dns-55f844cf75-pb59n\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.158287 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg6nq\" (UniqueName: \"kubernetes.io/projected/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-kube-api-access-vg6nq\") pod \"neutron-69b59f8dd-sckpz\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.158587 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-httpd-config\") pod \"neutron-69b59f8dd-sckpz\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.158655 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-combined-ca-bundle\") pod \"neutron-69b59f8dd-sckpz\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.158692 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-ovndb-tls-certs\") pod \"neutron-69b59f8dd-sckpz\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.158724 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-config\") pod \"neutron-69b59f8dd-sckpz\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.176335 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.264479 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-httpd-config\") pod \"neutron-69b59f8dd-sckpz\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.264605 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-combined-ca-bundle\") pod \"neutron-69b59f8dd-sckpz\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.264656 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-ovndb-tls-certs\") pod \"neutron-69b59f8dd-sckpz\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.264697 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-config\") pod \"neutron-69b59f8dd-sckpz\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.264745 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg6nq\" (UniqueName: \"kubernetes.io/projected/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-kube-api-access-vg6nq\") pod \"neutron-69b59f8dd-sckpz\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.273526 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-combined-ca-bundle\") pod \"neutron-69b59f8dd-sckpz\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.278304 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-httpd-config\") pod \"neutron-69b59f8dd-sckpz\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.292884 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg6nq\" (UniqueName: \"kubernetes.io/projected/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-kube-api-access-vg6nq\") pod \"neutron-69b59f8dd-sckpz\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.293601 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-ovndb-tls-certs\") pod \"neutron-69b59f8dd-sckpz\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.302329 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-config\") pod \"neutron-69b59f8dd-sckpz\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.486570 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.734460 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-pb59n"] Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.745514 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f","Type":"ContainerStarted","Data":"f5d9e63e248172bb123c2119799b48c42c47dc27abf1ccde1c22c8db92974276"} Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.775976 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68b68c8c84-qt25s" event={"ID":"23c89695-46b0-49d8-ab09-eb002d3cb94e","Type":"ContainerStarted","Data":"cfd7b4d9bffa269ff6f374499e736905fa75620e15e3e81a5e29e451252c437b"} Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.810893 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"34a77ce6-1705-45d6-af7f-8100e13a6d3b","Type":"ContainerStarted","Data":"a463adb758a277f6ea665085b981251f1e9da331319b2d1f8de125625611894b"} Dec 03 06:46:37 crc kubenswrapper[4818]: I1203 06:46:37.823047 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574dc6486d-nvkzv" event={"ID":"bd891b60-48cc-49f1-bb61-9e58c712816e","Type":"ContainerStarted","Data":"b75008d167e0556d0fd21ae4fa3d86d4b00028e3a7b752c00c1288d30868b6b8"} Dec 03 06:46:38 crc kubenswrapper[4818]: I1203 06:46:38.061410 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:46:38 crc kubenswrapper[4818]: I1203 06:46:38.207333 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69b59f8dd-sckpz"] Dec 03 06:46:38 crc kubenswrapper[4818]: W1203 06:46:38.226410 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff5cfbd5_5815_4ab8_8d5e_c7afbea33a9b.slice/crio-88518310e0da07a729cea5e9ee0a25e8db48bad877849ae2d1daf5d24465f428 WatchSource:0}: Error finding container 88518310e0da07a729cea5e9ee0a25e8db48bad877849ae2d1daf5d24465f428: Status 404 returned error can't find the container with id 88518310e0da07a729cea5e9ee0a25e8db48bad877849ae2d1daf5d24465f428 Dec 03 06:46:38 crc kubenswrapper[4818]: I1203 06:46:38.861531 4818 generic.go:334] "Generic (PLEG): container finished" podID="c23baebd-4d73-4972-b9eb-5c573e859977" containerID="a20b9ba01523bf083f62485b116ef7085bb29d10e0791cfce5c65d1bbc2107f9" exitCode=0 Dec 03 06:46:38 crc kubenswrapper[4818]: I1203 06:46:38.861913 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lqzrp" event={"ID":"c23baebd-4d73-4972-b9eb-5c573e859977","Type":"ContainerDied","Data":"a20b9ba01523bf083f62485b116ef7085bb29d10e0791cfce5c65d1bbc2107f9"} Dec 03 06:46:38 crc kubenswrapper[4818]: I1203 06:46:38.864783 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f","Type":"ContainerStarted","Data":"d741ce548f75e3da7cf0a8bd816e096a18f0c6cec4097ed3fe4b1cb7da83e0aa"} Dec 03 06:46:38 crc kubenswrapper[4818]: I1203 06:46:38.871450 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"34a77ce6-1705-45d6-af7f-8100e13a6d3b","Type":"ContainerStarted","Data":"31516935f93bba37172720180afa57d9aa0dbfa9449a6a0543bde59e2a83b036"} Dec 03 06:46:38 crc kubenswrapper[4818]: I1203 06:46:38.878108 4818 generic.go:334] "Generic (PLEG): container finished" podID="8cafb812-b94e-45b2-9f13-9a2be0599382" containerID="854ceb71d2a1feebf90c5ab71c605729b13e2996f08d6a534983921540014907" exitCode=0 Dec 03 06:46:38 crc kubenswrapper[4818]: I1203 06:46:38.878376 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-pb59n" event={"ID":"8cafb812-b94e-45b2-9f13-9a2be0599382","Type":"ContainerDied","Data":"854ceb71d2a1feebf90c5ab71c605729b13e2996f08d6a534983921540014907"} Dec 03 06:46:38 crc kubenswrapper[4818]: I1203 06:46:38.878464 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-pb59n" event={"ID":"8cafb812-b94e-45b2-9f13-9a2be0599382","Type":"ContainerStarted","Data":"5c93308997b1e2bfeb4f6e76ba5f51d6c08c50439e51abbc3bc8c0ecb8a3e200"} Dec 03 06:46:38 crc kubenswrapper[4818]: I1203 06:46:38.898174 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b59f8dd-sckpz" event={"ID":"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b","Type":"ContainerStarted","Data":"43fce69ae082d44a241742d1be5d81cb21889f23487c23b323a4b77d499b7453"} Dec 03 06:46:38 crc kubenswrapper[4818]: I1203 06:46:38.898231 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b59f8dd-sckpz" event={"ID":"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b","Type":"ContainerStarted","Data":"88518310e0da07a729cea5e9ee0a25e8db48bad877849ae2d1daf5d24465f428"} Dec 03 06:46:38 crc kubenswrapper[4818]: I1203 06:46:38.944244 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.944220445 podStartE2EDuration="5.944220445s" podCreationTimestamp="2025-12-03 06:46:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:46:38.93919068 +0000 UTC m=+1156.630799432" watchObservedRunningTime="2025-12-03 06:46:38.944220445 +0000 UTC m=+1156.635829197" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.693764 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=19.693726718 podStartE2EDuration="19.693726718s" podCreationTimestamp="2025-12-03 06:46:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:46:38.986678065 +0000 UTC m=+1156.678286827" watchObservedRunningTime="2025-12-03 06:46:39.693726718 +0000 UTC m=+1157.385335470" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.706467 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-77cd7f4557-lnszf"] Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.707978 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.709787 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.711464 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.727433 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77cd7f4557-lnszf"] Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.853872 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-internal-tls-certs\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.853946 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-httpd-config\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.854072 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-combined-ca-bundle\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.854112 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-config\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.854148 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4qdk\" (UniqueName: \"kubernetes.io/projected/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-kube-api-access-n4qdk\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.854177 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-ovndb-tls-certs\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.854194 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-public-tls-certs\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.916584 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-pb59n" event={"ID":"8cafb812-b94e-45b2-9f13-9a2be0599382","Type":"ContainerStarted","Data":"e7b9f5c9eaf99c43a223dbb69a7170a6b6cdb9b5ea6ff0763179bc3f0fbd12ef"} Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.919326 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b59f8dd-sckpz" event={"ID":"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b","Type":"ContainerStarted","Data":"45c9fba1cd213a75cebfe1bfc9c9756645c2b5a012d0442c0ffee17a3bc40bab"} Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.951615 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-69b59f8dd-sckpz" podStartSLOduration=3.951595555 podStartE2EDuration="3.951595555s" podCreationTimestamp="2025-12-03 06:46:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:46:39.9432892 +0000 UTC m=+1157.634897962" watchObservedRunningTime="2025-12-03 06:46:39.951595555 +0000 UTC m=+1157.643204307" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.955617 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-combined-ca-bundle\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.955663 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-config\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.955710 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4qdk\" (UniqueName: \"kubernetes.io/projected/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-kube-api-access-n4qdk\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.955729 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-ovndb-tls-certs\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.955753 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-public-tls-certs\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.955832 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-internal-tls-certs\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.955903 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-httpd-config\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.962679 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-httpd-config\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.967156 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-internal-tls-certs\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.968534 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-ovndb-tls-certs\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.976672 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-combined-ca-bundle\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.983834 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-config\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.985355 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4qdk\" (UniqueName: \"kubernetes.io/projected/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-kube-api-access-n4qdk\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:39 crc kubenswrapper[4818]: I1203 06:46:39.988159 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcf80b5c-a427-4bcb-a60f-51b730b41cf2-public-tls-certs\") pod \"neutron-77cd7f4557-lnszf\" (UID: \"bcf80b5c-a427-4bcb-a60f-51b730b41cf2\") " pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.023374 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.459580 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.576539 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-scripts\") pod \"c23baebd-4d73-4972-b9eb-5c573e859977\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.576975 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-config-data\") pod \"c23baebd-4d73-4972-b9eb-5c573e859977\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.577018 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c23baebd-4d73-4972-b9eb-5c573e859977-logs\") pod \"c23baebd-4d73-4972-b9eb-5c573e859977\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.577053 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-combined-ca-bundle\") pod \"c23baebd-4d73-4972-b9eb-5c573e859977\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.577108 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zdmf\" (UniqueName: \"kubernetes.io/projected/c23baebd-4d73-4972-b9eb-5c573e859977-kube-api-access-8zdmf\") pod \"c23baebd-4d73-4972-b9eb-5c573e859977\" (UID: \"c23baebd-4d73-4972-b9eb-5c573e859977\") " Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.579190 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c23baebd-4d73-4972-b9eb-5c573e859977-logs" (OuterVolumeSpecName: "logs") pod "c23baebd-4d73-4972-b9eb-5c573e859977" (UID: "c23baebd-4d73-4972-b9eb-5c573e859977"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.588990 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-scripts" (OuterVolumeSpecName: "scripts") pod "c23baebd-4d73-4972-b9eb-5c573e859977" (UID: "c23baebd-4d73-4972-b9eb-5c573e859977"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.598982 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c23baebd-4d73-4972-b9eb-5c573e859977-kube-api-access-8zdmf" (OuterVolumeSpecName: "kube-api-access-8zdmf") pod "c23baebd-4d73-4972-b9eb-5c573e859977" (UID: "c23baebd-4d73-4972-b9eb-5c573e859977"). InnerVolumeSpecName "kube-api-access-8zdmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.628947 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c23baebd-4d73-4972-b9eb-5c573e859977" (UID: "c23baebd-4d73-4972-b9eb-5c573e859977"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.633679 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-config-data" (OuterVolumeSpecName: "config-data") pod "c23baebd-4d73-4972-b9eb-5c573e859977" (UID: "c23baebd-4d73-4972-b9eb-5c573e859977"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.678844 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zdmf\" (UniqueName: \"kubernetes.io/projected/c23baebd-4d73-4972-b9eb-5c573e859977-kube-api-access-8zdmf\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.678880 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.678889 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.678900 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c23baebd-4d73-4972-b9eb-5c573e859977-logs\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.678908 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c23baebd-4d73-4972-b9eb-5c573e859977-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.891913 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77cd7f4557-lnszf"] Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.967126 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lqzrp" event={"ID":"c23baebd-4d73-4972-b9eb-5c573e859977","Type":"ContainerDied","Data":"e540c77a9433517242e3f821216550b8dc2f195a78e8c037fff11108cca47a4e"} Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.967164 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e540c77a9433517242e3f821216550b8dc2f195a78e8c037fff11108cca47a4e" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.967201 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lqzrp" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.976394 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77cd7f4557-lnszf" event={"ID":"bcf80b5c-a427-4bcb-a60f-51b730b41cf2","Type":"ContainerStarted","Data":"c198a87e45776bf81527f38213259e868fa26d9383107d5c005a9b1c64e99bb4"} Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.980001 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.980023 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.993014 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 06:46:40 crc kubenswrapper[4818]: I1203 06:46:40.993055 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.027674 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-pb59n" podStartSLOduration=5.027653854 podStartE2EDuration="5.027653854s" podCreationTimestamp="2025-12-03 06:46:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:46:41.019918442 +0000 UTC m=+1158.711527194" watchObservedRunningTime="2025-12-03 06:46:41.027653854 +0000 UTC m=+1158.719262606" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.027957 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-748bc8fb48-scr66"] Dec 03 06:46:41 crc kubenswrapper[4818]: E1203 06:46:41.031533 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c23baebd-4d73-4972-b9eb-5c573e859977" containerName="placement-db-sync" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.035110 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c23baebd-4d73-4972-b9eb-5c573e859977" containerName="placement-db-sync" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.035503 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="c23baebd-4d73-4972-b9eb-5c573e859977" containerName="placement-db-sync" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.036422 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.039085 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.039442 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.039608 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-n4gmb" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.040913 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.054630 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.055132 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.109379 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.116152 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-748bc8fb48-scr66"] Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.216255 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-combined-ca-bundle\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.216421 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-logs\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.216511 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-public-tls-certs\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.216535 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-internal-tls-certs\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.216562 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-config-data\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.216650 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8cm7\" (UniqueName: \"kubernetes.io/projected/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-kube-api-access-f8cm7\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.216688 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-scripts\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.323931 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-public-tls-certs\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.324010 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-internal-tls-certs\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.324048 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-config-data\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.324178 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8cm7\" (UniqueName: \"kubernetes.io/projected/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-kube-api-access-f8cm7\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.324227 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-scripts\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.324437 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-combined-ca-bundle\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.324512 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-logs\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.325390 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-logs\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.328049 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-public-tls-certs\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.329367 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-config-data\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.334767 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-combined-ca-bundle\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.335139 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-internal-tls-certs\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.352264 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-scripts\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.374577 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8cm7\" (UniqueName: \"kubernetes.io/projected/e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff-kube-api-access-f8cm7\") pod \"placement-748bc8fb48-scr66\" (UID: \"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff\") " pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:41 crc kubenswrapper[4818]: I1203 06:46:41.668098 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:42 crc kubenswrapper[4818]: I1203 06:46:42.002221 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77cd7f4557-lnszf" event={"ID":"bcf80b5c-a427-4bcb-a60f-51b730b41cf2","Type":"ContainerStarted","Data":"b86be19ead52deebb8c1cdce2983f2384c65ea9101a281439ac8278064f04418"} Dec 03 06:46:42 crc kubenswrapper[4818]: I1203 06:46:42.002799 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 06:46:42 crc kubenswrapper[4818]: I1203 06:46:42.003147 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 06:46:42 crc kubenswrapper[4818]: I1203 06:46:42.199401 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-748bc8fb48-scr66"] Dec 03 06:46:42 crc kubenswrapper[4818]: W1203 06:46:42.210025 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8fdfb1c_8a82_41cb_a8e0_58e83d5bc8ff.slice/crio-954ea35a3ef5eccff876f640e98bd7da7bab6cf1b4e141fb018518d25d8d33f8 WatchSource:0}: Error finding container 954ea35a3ef5eccff876f640e98bd7da7bab6cf1b4e141fb018518d25d8d33f8: Status 404 returned error can't find the container with id 954ea35a3ef5eccff876f640e98bd7da7bab6cf1b4e141fb018518d25d8d33f8 Dec 03 06:46:43 crc kubenswrapper[4818]: I1203 06:46:43.129369 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-748bc8fb48-scr66" event={"ID":"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff","Type":"ContainerStarted","Data":"3a95a61295b7d5e90c47a660a20cc74c1b072057fc64d98e1f66f4bd3596a438"} Dec 03 06:46:43 crc kubenswrapper[4818]: I1203 06:46:43.129756 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-748bc8fb48-scr66" event={"ID":"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff","Type":"ContainerStarted","Data":"954ea35a3ef5eccff876f640e98bd7da7bab6cf1b4e141fb018518d25d8d33f8"} Dec 03 06:46:43 crc kubenswrapper[4818]: I1203 06:46:43.138489 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77cd7f4557-lnszf" event={"ID":"bcf80b5c-a427-4bcb-a60f-51b730b41cf2","Type":"ContainerStarted","Data":"2e95a99dab57fe0d28aa13133d4d5ada873265eff25f2b128e21062f694fcffb"} Dec 03 06:46:43 crc kubenswrapper[4818]: I1203 06:46:43.138865 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:46:43 crc kubenswrapper[4818]: I1203 06:46:43.143893 4818 generic.go:334] "Generic (PLEG): container finished" podID="56e81897-8229-4c21-9beb-a7b0941edf03" containerID="19a34de912c8415893becf259c87b4bece98c0984dc27039b91e46ed8aeb25b5" exitCode=0 Dec 03 06:46:43 crc kubenswrapper[4818]: I1203 06:46:43.144048 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v8jgr" event={"ID":"56e81897-8229-4c21-9beb-a7b0941edf03","Type":"ContainerDied","Data":"19a34de912c8415893becf259c87b4bece98c0984dc27039b91e46ed8aeb25b5"} Dec 03 06:46:43 crc kubenswrapper[4818]: I1203 06:46:43.167745 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-77cd7f4557-lnszf" podStartSLOduration=4.167722591 podStartE2EDuration="4.167722591s" podCreationTimestamp="2025-12-03 06:46:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:46:43.161329833 +0000 UTC m=+1160.852938605" watchObservedRunningTime="2025-12-03 06:46:43.167722591 +0000 UTC m=+1160.859331343" Dec 03 06:46:43 crc kubenswrapper[4818]: I1203 06:46:43.302094 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:46:43 crc kubenswrapper[4818]: I1203 06:46:43.302497 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:46:43 crc kubenswrapper[4818]: I1203 06:46:43.846224 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 06:46:43 crc kubenswrapper[4818]: I1203 06:46:43.848081 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 06:46:43 crc kubenswrapper[4818]: I1203 06:46:43.894847 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 06:46:43 crc kubenswrapper[4818]: I1203 06:46:43.907027 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 06:46:44 crc kubenswrapper[4818]: I1203 06:46:44.153728 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:46:44 crc kubenswrapper[4818]: I1203 06:46:44.153754 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:46:44 crc kubenswrapper[4818]: I1203 06:46:44.154398 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 06:46:44 crc kubenswrapper[4818]: I1203 06:46:44.164500 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 06:46:45 crc kubenswrapper[4818]: I1203 06:46:45.684390 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:45 crc kubenswrapper[4818]: I1203 06:46:45.684950 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:46:45 crc kubenswrapper[4818]: I1203 06:46:45.750190 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 06:46:45 crc kubenswrapper[4818]: I1203 06:46:45.750292 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:46:45 crc kubenswrapper[4818]: I1203 06:46:45.752235 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 06:46:45 crc kubenswrapper[4818]: I1203 06:46:45.816113 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:45 crc kubenswrapper[4818]: I1203 06:46:45.816342 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:46:46 crc kubenswrapper[4818]: I1203 06:46:46.189973 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:46:46 crc kubenswrapper[4818]: I1203 06:46:46.189996 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:46:46 crc kubenswrapper[4818]: I1203 06:46:46.201920 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:46:46 crc kubenswrapper[4818]: I1203 06:46:46.558701 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 06:46:46 crc kubenswrapper[4818]: I1203 06:46:46.748773 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 06:46:47 crc kubenswrapper[4818]: I1203 06:46:47.178484 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:46:47 crc kubenswrapper[4818]: I1203 06:46:47.254654 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-9kv6h"] Dec 03 06:46:47 crc kubenswrapper[4818]: I1203 06:46:47.254934 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" podUID="54cbca7f-1646-4eaa-a61b-6094b8506572" containerName="dnsmasq-dns" containerID="cri-o://8dd11ab6b6e7e9a1512584a33ea49bca6d1ef864366536d8dc3ad47ffb615895" gracePeriod=10 Dec 03 06:46:48 crc kubenswrapper[4818]: I1203 06:46:48.208375 4818 generic.go:334] "Generic (PLEG): container finished" podID="54cbca7f-1646-4eaa-a61b-6094b8506572" containerID="8dd11ab6b6e7e9a1512584a33ea49bca6d1ef864366536d8dc3ad47ffb615895" exitCode=0 Dec 03 06:46:48 crc kubenswrapper[4818]: I1203 06:46:48.208565 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" event={"ID":"54cbca7f-1646-4eaa-a61b-6094b8506572","Type":"ContainerDied","Data":"8dd11ab6b6e7e9a1512584a33ea49bca6d1ef864366536d8dc3ad47ffb615895"} Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.097013 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.159111 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-config-data\") pod \"56e81897-8229-4c21-9beb-a7b0941edf03\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.159190 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-combined-ca-bundle\") pod \"56e81897-8229-4c21-9beb-a7b0941edf03\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.159258 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-credential-keys\") pod \"56e81897-8229-4c21-9beb-a7b0941edf03\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.159286 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-fernet-keys\") pod \"56e81897-8229-4c21-9beb-a7b0941edf03\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.159322 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-scripts\") pod \"56e81897-8229-4c21-9beb-a7b0941edf03\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.159400 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srhgb\" (UniqueName: \"kubernetes.io/projected/56e81897-8229-4c21-9beb-a7b0941edf03-kube-api-access-srhgb\") pod \"56e81897-8229-4c21-9beb-a7b0941edf03\" (UID: \"56e81897-8229-4c21-9beb-a7b0941edf03\") " Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.166960 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56e81897-8229-4c21-9beb-a7b0941edf03-kube-api-access-srhgb" (OuterVolumeSpecName: "kube-api-access-srhgb") pod "56e81897-8229-4c21-9beb-a7b0941edf03" (UID: "56e81897-8229-4c21-9beb-a7b0941edf03"). InnerVolumeSpecName "kube-api-access-srhgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.188750 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "56e81897-8229-4c21-9beb-a7b0941edf03" (UID: "56e81897-8229-4c21-9beb-a7b0941edf03"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.193027 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "56e81897-8229-4c21-9beb-a7b0941edf03" (UID: "56e81897-8229-4c21-9beb-a7b0941edf03"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.199510 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-scripts" (OuterVolumeSpecName: "scripts") pod "56e81897-8229-4c21-9beb-a7b0941edf03" (UID: "56e81897-8229-4c21-9beb-a7b0941edf03"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.237279 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-config-data" (OuterVolumeSpecName: "config-data") pod "56e81897-8229-4c21-9beb-a7b0941edf03" (UID: "56e81897-8229-4c21-9beb-a7b0941edf03"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.263370 4818 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.263406 4818 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.263420 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.263432 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srhgb\" (UniqueName: \"kubernetes.io/projected/56e81897-8229-4c21-9beb-a7b0941edf03-kube-api-access-srhgb\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.263447 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.278159 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v8jgr" event={"ID":"56e81897-8229-4c21-9beb-a7b0941edf03","Type":"ContainerDied","Data":"dd918ad3067671c1127691fb2afa4ae2c4b58a5c4c3cc1040b4cfe018533952e"} Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.278522 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd918ad3067671c1127691fb2afa4ae2c4b58a5c4c3cc1040b4cfe018533952e" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.278430 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v8jgr" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.300958 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56e81897-8229-4c21-9beb-a7b0941edf03" (UID: "56e81897-8229-4c21-9beb-a7b0941edf03"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.329371 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.372930 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56e81897-8229-4c21-9beb-a7b0941edf03-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.474242 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-config\") pod \"54cbca7f-1646-4eaa-a61b-6094b8506572\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.474360 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-dns-swift-storage-0\") pod \"54cbca7f-1646-4eaa-a61b-6094b8506572\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.474471 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct4xm\" (UniqueName: \"kubernetes.io/projected/54cbca7f-1646-4eaa-a61b-6094b8506572-kube-api-access-ct4xm\") pod \"54cbca7f-1646-4eaa-a61b-6094b8506572\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.474501 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-ovsdbserver-nb\") pod \"54cbca7f-1646-4eaa-a61b-6094b8506572\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.474535 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-dns-svc\") pod \"54cbca7f-1646-4eaa-a61b-6094b8506572\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.474595 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-ovsdbserver-sb\") pod \"54cbca7f-1646-4eaa-a61b-6094b8506572\" (UID: \"54cbca7f-1646-4eaa-a61b-6094b8506572\") " Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.480969 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54cbca7f-1646-4eaa-a61b-6094b8506572-kube-api-access-ct4xm" (OuterVolumeSpecName: "kube-api-access-ct4xm") pod "54cbca7f-1646-4eaa-a61b-6094b8506572" (UID: "54cbca7f-1646-4eaa-a61b-6094b8506572"). InnerVolumeSpecName "kube-api-access-ct4xm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.540694 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "54cbca7f-1646-4eaa-a61b-6094b8506572" (UID: "54cbca7f-1646-4eaa-a61b-6094b8506572"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.545256 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "54cbca7f-1646-4eaa-a61b-6094b8506572" (UID: "54cbca7f-1646-4eaa-a61b-6094b8506572"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.546380 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "54cbca7f-1646-4eaa-a61b-6094b8506572" (UID: "54cbca7f-1646-4eaa-a61b-6094b8506572"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.553216 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "54cbca7f-1646-4eaa-a61b-6094b8506572" (UID: "54cbca7f-1646-4eaa-a61b-6094b8506572"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.564448 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-config" (OuterVolumeSpecName: "config") pod "54cbca7f-1646-4eaa-a61b-6094b8506572" (UID: "54cbca7f-1646-4eaa-a61b-6094b8506572"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.576782 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.576841 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.576856 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.576869 4818 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.576882 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct4xm\" (UniqueName: \"kubernetes.io/projected/54cbca7f-1646-4eaa-a61b-6094b8506572-kube-api-access-ct4xm\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:49 crc kubenswrapper[4818]: I1203 06:46:49.576893 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/54cbca7f-1646-4eaa-a61b-6094b8506572-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.209085 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-75b5b56458-pbq6h"] Dec 03 06:46:50 crc kubenswrapper[4818]: E1203 06:46:50.209745 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cbca7f-1646-4eaa-a61b-6094b8506572" containerName="dnsmasq-dns" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.209763 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cbca7f-1646-4eaa-a61b-6094b8506572" containerName="dnsmasq-dns" Dec 03 06:46:50 crc kubenswrapper[4818]: E1203 06:46:50.209786 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56e81897-8229-4c21-9beb-a7b0941edf03" containerName="keystone-bootstrap" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.209792 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="56e81897-8229-4c21-9beb-a7b0941edf03" containerName="keystone-bootstrap" Dec 03 06:46:50 crc kubenswrapper[4818]: E1203 06:46:50.210885 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cbca7f-1646-4eaa-a61b-6094b8506572" containerName="init" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.210900 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cbca7f-1646-4eaa-a61b-6094b8506572" containerName="init" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.211071 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="56e81897-8229-4c21-9beb-a7b0941edf03" containerName="keystone-bootstrap" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.211099 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cbca7f-1646-4eaa-a61b-6094b8506572" containerName="dnsmasq-dns" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.211688 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.214003 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ggd6v" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.214123 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.214331 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.217844 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.218296 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.220913 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.226902 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-75b5b56458-pbq6h"] Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.311866 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-748bc8fb48-scr66" event={"ID":"e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff","Type":"ContainerStarted","Data":"c049daf8e24e16ce3f806e79bc99c5fbd7d5eb24c1a194b86a938d63d5f162ea"} Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.312872 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-credential-keys\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.312910 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-scripts\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.312920 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.312936 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-internal-tls-certs\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.312989 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlbfm\" (UniqueName: \"kubernetes.io/projected/29098cb1-d622-4ee2-ad26-f8d51eec41f4-kube-api-access-zlbfm\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.313045 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-combined-ca-bundle\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.313076 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-config-data\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.313153 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-public-tls-certs\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.313190 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-fernet-keys\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.313219 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.328029 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7kbvz" event={"ID":"982110da-ac09-4df8-9729-77284ed7539b","Type":"ContainerStarted","Data":"8ff182159e975cf1a6bd5f92183c3e16693a90e305a4a18dca48b23710bd8cee"} Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.337759 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"575fe180-2a23-48e9-b5cf-b90e83528e5e","Type":"ContainerStarted","Data":"693c121be8011128371d293e5fc85539a1e5af22221f56702ec76f8eb5db0714"} Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.339692 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" event={"ID":"54cbca7f-1646-4eaa-a61b-6094b8506572","Type":"ContainerDied","Data":"3fa180d4d46f7e5941905882704b9e463c75c28082bfd56e0a362d8b4d8615b9"} Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.339747 4818 scope.go:117] "RemoveContainer" containerID="8dd11ab6b6e7e9a1512584a33ea49bca6d1ef864366536d8dc3ad47ffb615895" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.339942 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-9kv6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.389028 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-748bc8fb48-scr66" podStartSLOduration=10.389006024 podStartE2EDuration="10.389006024s" podCreationTimestamp="2025-12-03 06:46:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:46:50.364632152 +0000 UTC m=+1168.056240914" watchObservedRunningTime="2025-12-03 06:46:50.389006024 +0000 UTC m=+1168.080614776" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.417774 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlbfm\" (UniqueName: \"kubernetes.io/projected/29098cb1-d622-4ee2-ad26-f8d51eec41f4-kube-api-access-zlbfm\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.418172 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-combined-ca-bundle\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.418297 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-config-data\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.418526 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-public-tls-certs\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.418675 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-fernet-keys\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.418804 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-credential-keys\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.418944 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-scripts\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.419065 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-internal-tls-certs\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.449613 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-7kbvz" podStartSLOduration=3.592684511 podStartE2EDuration="45.449592382s" podCreationTimestamp="2025-12-03 06:46:05 +0000 UTC" firstStartedPulling="2025-12-03 06:46:07.394726899 +0000 UTC m=+1125.086335651" lastFinishedPulling="2025-12-03 06:46:49.25163477 +0000 UTC m=+1166.943243522" observedRunningTime="2025-12-03 06:46:50.432314675 +0000 UTC m=+1168.123923427" watchObservedRunningTime="2025-12-03 06:46:50.449592382 +0000 UTC m=+1168.141201134" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.464223 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-internal-tls-certs\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.466994 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-public-tls-certs\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.467278 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-combined-ca-bundle\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.467441 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-fernet-keys\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.492027 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-config-data\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.500500 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-scripts\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.501738 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlbfm\" (UniqueName: \"kubernetes.io/projected/29098cb1-d622-4ee2-ad26-f8d51eec41f4-kube-api-access-zlbfm\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.516276 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-9kv6h"] Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.530622 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/29098cb1-d622-4ee2-ad26-f8d51eec41f4-credential-keys\") pod \"keystone-75b5b56458-pbq6h\" (UID: \"29098cb1-d622-4ee2-ad26-f8d51eec41f4\") " pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.554617 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.575191 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-9kv6h"] Dec 03 06:46:50 crc kubenswrapper[4818]: I1203 06:46:50.753963 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54cbca7f-1646-4eaa-a61b-6094b8506572" path="/var/lib/kubelet/pods/54cbca7f-1646-4eaa-a61b-6094b8506572/volumes" Dec 03 06:46:51 crc kubenswrapper[4818]: I1203 06:46:51.382871 4818 scope.go:117] "RemoveContainer" containerID="22290b7ddabe8ab3beed9e02937c9b2e354cb3bc3c99f0fcf90aa565869c3505" Dec 03 06:46:51 crc kubenswrapper[4818]: I1203 06:46:51.961834 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-75b5b56458-pbq6h"] Dec 03 06:46:52 crc kubenswrapper[4818]: I1203 06:46:52.171466 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:46:52 crc kubenswrapper[4818]: I1203 06:46:52.365113 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-75b5b56458-pbq6h" event={"ID":"29098cb1-d622-4ee2-ad26-f8d51eec41f4","Type":"ContainerStarted","Data":"3b3e3131c3287ec2f05d7924dcdbdf94d7d82081aa1d85ff9e993af3ee784738"} Dec 03 06:46:52 crc kubenswrapper[4818]: I1203 06:46:52.365431 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-75b5b56458-pbq6h" event={"ID":"29098cb1-d622-4ee2-ad26-f8d51eec41f4","Type":"ContainerStarted","Data":"02d8ec4ef386bc1826446dbd9c416be501a08bbeb13a15926955d71c26ba03b7"} Dec 03 06:46:52 crc kubenswrapper[4818]: I1203 06:46:52.365448 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:46:52 crc kubenswrapper[4818]: I1203 06:46:52.393903 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-75b5b56458-pbq6h" podStartSLOduration=2.393887477 podStartE2EDuration="2.393887477s" podCreationTimestamp="2025-12-03 06:46:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:46:52.388468934 +0000 UTC m=+1170.080077686" watchObservedRunningTime="2025-12-03 06:46:52.393887477 +0000 UTC m=+1170.085496229" Dec 03 06:46:53 crc kubenswrapper[4818]: I1203 06:46:53.401874 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-978k5" event={"ID":"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3","Type":"ContainerStarted","Data":"2eec35e25d66cc807661730864b68b790a563c984227ff43593f4242c9b2ed83"} Dec 03 06:46:53 crc kubenswrapper[4818]: I1203 06:46:53.428332 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-978k5" podStartSLOduration=3.838840426 podStartE2EDuration="48.428316163s" podCreationTimestamp="2025-12-03 06:46:05 +0000 UTC" firstStartedPulling="2025-12-03 06:46:06.834148777 +0000 UTC m=+1124.525757519" lastFinishedPulling="2025-12-03 06:46:51.423624504 +0000 UTC m=+1169.115233256" observedRunningTime="2025-12-03 06:46:53.427190286 +0000 UTC m=+1171.118799038" watchObservedRunningTime="2025-12-03 06:46:53.428316163 +0000 UTC m=+1171.119924905" Dec 03 06:46:54 crc kubenswrapper[4818]: I1203 06:46:54.411274 4818 generic.go:334] "Generic (PLEG): container finished" podID="982110da-ac09-4df8-9729-77284ed7539b" containerID="8ff182159e975cf1a6bd5f92183c3e16693a90e305a4a18dca48b23710bd8cee" exitCode=0 Dec 03 06:46:54 crc kubenswrapper[4818]: I1203 06:46:54.411622 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7kbvz" event={"ID":"982110da-ac09-4df8-9729-77284ed7539b","Type":"ContainerDied","Data":"8ff182159e975cf1a6bd5f92183c3e16693a90e305a4a18dca48b23710bd8cee"} Dec 03 06:46:55 crc kubenswrapper[4818]: I1203 06:46:55.704203 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-574dc6486d-nvkzv" podUID="bd891b60-48cc-49f1-bb61-9e58c712816e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 03 06:46:55 crc kubenswrapper[4818]: I1203 06:46:55.813438 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7kbvz" Dec 03 06:46:55 crc kubenswrapper[4818]: I1203 06:46:55.818377 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68b68c8c84-qt25s" podUID="23c89695-46b0-49d8-ab09-eb002d3cb94e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 03 06:46:55 crc kubenswrapper[4818]: I1203 06:46:55.972411 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/982110da-ac09-4df8-9729-77284ed7539b-db-sync-config-data\") pod \"982110da-ac09-4df8-9729-77284ed7539b\" (UID: \"982110da-ac09-4df8-9729-77284ed7539b\") " Dec 03 06:46:55 crc kubenswrapper[4818]: I1203 06:46:55.973952 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/982110da-ac09-4df8-9729-77284ed7539b-combined-ca-bundle\") pod \"982110da-ac09-4df8-9729-77284ed7539b\" (UID: \"982110da-ac09-4df8-9729-77284ed7539b\") " Dec 03 06:46:55 crc kubenswrapper[4818]: I1203 06:46:55.974123 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwqv8\" (UniqueName: \"kubernetes.io/projected/982110da-ac09-4df8-9729-77284ed7539b-kube-api-access-vwqv8\") pod \"982110da-ac09-4df8-9729-77284ed7539b\" (UID: \"982110da-ac09-4df8-9729-77284ed7539b\") " Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:55.997851 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/982110da-ac09-4df8-9729-77284ed7539b-kube-api-access-vwqv8" (OuterVolumeSpecName: "kube-api-access-vwqv8") pod "982110da-ac09-4df8-9729-77284ed7539b" (UID: "982110da-ac09-4df8-9729-77284ed7539b"). InnerVolumeSpecName "kube-api-access-vwqv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.005766 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/982110da-ac09-4df8-9729-77284ed7539b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "982110da-ac09-4df8-9729-77284ed7539b" (UID: "982110da-ac09-4df8-9729-77284ed7539b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.009414 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/982110da-ac09-4df8-9729-77284ed7539b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "982110da-ac09-4df8-9729-77284ed7539b" (UID: "982110da-ac09-4df8-9729-77284ed7539b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.076937 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/982110da-ac09-4df8-9729-77284ed7539b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.076977 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwqv8\" (UniqueName: \"kubernetes.io/projected/982110da-ac09-4df8-9729-77284ed7539b-kube-api-access-vwqv8\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.076992 4818 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/982110da-ac09-4df8-9729-77284ed7539b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.433082 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-7kbvz" event={"ID":"982110da-ac09-4df8-9729-77284ed7539b","Type":"ContainerDied","Data":"b870c49ea2ff79b6648f1f60c7f0e0f53ec7381310232bb5cc9985375a4527ab"} Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.433120 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b870c49ea2ff79b6648f1f60c7f0e0f53ec7381310232bb5cc9985375a4527ab" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.433133 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-7kbvz" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.627015 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-554df66f87-jscgb"] Dec 03 06:46:56 crc kubenswrapper[4818]: E1203 06:46:56.627742 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="982110da-ac09-4df8-9729-77284ed7539b" containerName="barbican-db-sync" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.627759 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="982110da-ac09-4df8-9729-77284ed7539b" containerName="barbican-db-sync" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.627972 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="982110da-ac09-4df8-9729-77284ed7539b" containerName="barbican-db-sync" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.628936 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-554df66f87-jscgb" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.636029 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.636281 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.636434 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-2n894" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.687824 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8e58804-c7d8-48ce-9db3-6c3ab56f1c13-combined-ca-bundle\") pod \"barbican-worker-554df66f87-jscgb\" (UID: \"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13\") " pod="openstack/barbican-worker-554df66f87-jscgb" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.692190 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8e58804-c7d8-48ce-9db3-6c3ab56f1c13-config-data-custom\") pod \"barbican-worker-554df66f87-jscgb\" (UID: \"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13\") " pod="openstack/barbican-worker-554df66f87-jscgb" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.692503 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8e58804-c7d8-48ce-9db3-6c3ab56f1c13-config-data\") pod \"barbican-worker-554df66f87-jscgb\" (UID: \"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13\") " pod="openstack/barbican-worker-554df66f87-jscgb" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.692621 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8e58804-c7d8-48ce-9db3-6c3ab56f1c13-logs\") pod \"barbican-worker-554df66f87-jscgb\" (UID: \"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13\") " pod="openstack/barbican-worker-554df66f87-jscgb" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.692747 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rgvz\" (UniqueName: \"kubernetes.io/projected/d8e58804-c7d8-48ce-9db3-6c3ab56f1c13-kube-api-access-7rgvz\") pod \"barbican-worker-554df66f87-jscgb\" (UID: \"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13\") " pod="openstack/barbican-worker-554df66f87-jscgb" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.696181 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7b4848f678-kfvpp"] Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.722411 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7b4848f678-kfvpp"] Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.722537 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.743843 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.794502 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rgvz\" (UniqueName: \"kubernetes.io/projected/d8e58804-c7d8-48ce-9db3-6c3ab56f1c13-kube-api-access-7rgvz\") pod \"barbican-worker-554df66f87-jscgb\" (UID: \"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13\") " pod="openstack/barbican-worker-554df66f87-jscgb" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.794575 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8e58804-c7d8-48ce-9db3-6c3ab56f1c13-combined-ca-bundle\") pod \"barbican-worker-554df66f87-jscgb\" (UID: \"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13\") " pod="openstack/barbican-worker-554df66f87-jscgb" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.794616 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8e58804-c7d8-48ce-9db3-6c3ab56f1c13-config-data-custom\") pod \"barbican-worker-554df66f87-jscgb\" (UID: \"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13\") " pod="openstack/barbican-worker-554df66f87-jscgb" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.794687 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d8d8f21-1f1f-4431-8079-ffad453ff3f1-combined-ca-bundle\") pod \"barbican-keystone-listener-7b4848f678-kfvpp\" (UID: \"6d8d8f21-1f1f-4431-8079-ffad453ff3f1\") " pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.794707 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6d8d8f21-1f1f-4431-8079-ffad453ff3f1-config-data-custom\") pod \"barbican-keystone-listener-7b4848f678-kfvpp\" (UID: \"6d8d8f21-1f1f-4431-8079-ffad453ff3f1\") " pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.794726 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d8d8f21-1f1f-4431-8079-ffad453ff3f1-logs\") pod \"barbican-keystone-listener-7b4848f678-kfvpp\" (UID: \"6d8d8f21-1f1f-4431-8079-ffad453ff3f1\") " pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.794775 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d8d8f21-1f1f-4431-8079-ffad453ff3f1-config-data\") pod \"barbican-keystone-listener-7b4848f678-kfvpp\" (UID: \"6d8d8f21-1f1f-4431-8079-ffad453ff3f1\") " pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.796677 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8e58804-c7d8-48ce-9db3-6c3ab56f1c13-config-data\") pod \"barbican-worker-554df66f87-jscgb\" (UID: \"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13\") " pod="openstack/barbican-worker-554df66f87-jscgb" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.796804 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpmwq\" (UniqueName: \"kubernetes.io/projected/6d8d8f21-1f1f-4431-8079-ffad453ff3f1-kube-api-access-kpmwq\") pod \"barbican-keystone-listener-7b4848f678-kfvpp\" (UID: \"6d8d8f21-1f1f-4431-8079-ffad453ff3f1\") " pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.805952 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8e58804-c7d8-48ce-9db3-6c3ab56f1c13-logs\") pod \"barbican-worker-554df66f87-jscgb\" (UID: \"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13\") " pod="openstack/barbican-worker-554df66f87-jscgb" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.806436 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8e58804-c7d8-48ce-9db3-6c3ab56f1c13-logs\") pod \"barbican-worker-554df66f87-jscgb\" (UID: \"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13\") " pod="openstack/barbican-worker-554df66f87-jscgb" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.824344 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-554df66f87-jscgb"] Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.829565 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8e58804-c7d8-48ce-9db3-6c3ab56f1c13-config-data\") pod \"barbican-worker-554df66f87-jscgb\" (UID: \"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13\") " pod="openstack/barbican-worker-554df66f87-jscgb" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.861705 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8e58804-c7d8-48ce-9db3-6c3ab56f1c13-combined-ca-bundle\") pod \"barbican-worker-554df66f87-jscgb\" (UID: \"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13\") " pod="openstack/barbican-worker-554df66f87-jscgb" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.871500 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8e58804-c7d8-48ce-9db3-6c3ab56f1c13-config-data-custom\") pod \"barbican-worker-554df66f87-jscgb\" (UID: \"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13\") " pod="openstack/barbican-worker-554df66f87-jscgb" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.871576 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-g9dwh"] Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.873532 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rgvz\" (UniqueName: \"kubernetes.io/projected/d8e58804-c7d8-48ce-9db3-6c3ab56f1c13-kube-api-access-7rgvz\") pod \"barbican-worker-554df66f87-jscgb\" (UID: \"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13\") " pod="openstack/barbican-worker-554df66f87-jscgb" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.886049 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-g9dwh"] Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.886184 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.935717 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpmwq\" (UniqueName: \"kubernetes.io/projected/6d8d8f21-1f1f-4431-8079-ffad453ff3f1-kube-api-access-kpmwq\") pod \"barbican-keystone-listener-7b4848f678-kfvpp\" (UID: \"6d8d8f21-1f1f-4431-8079-ffad453ff3f1\") " pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.935764 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.935799 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.935871 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2r85\" (UniqueName: \"kubernetes.io/projected/add5819a-c589-4b78-9132-1a8166206a9e-kube-api-access-w2r85\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.935935 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-dns-svc\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.935955 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.935987 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d8d8f21-1f1f-4431-8079-ffad453ff3f1-combined-ca-bundle\") pod \"barbican-keystone-listener-7b4848f678-kfvpp\" (UID: \"6d8d8f21-1f1f-4431-8079-ffad453ff3f1\") " pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.936018 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6d8d8f21-1f1f-4431-8079-ffad453ff3f1-config-data-custom\") pod \"barbican-keystone-listener-7b4848f678-kfvpp\" (UID: \"6d8d8f21-1f1f-4431-8079-ffad453ff3f1\") " pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.936043 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d8d8f21-1f1f-4431-8079-ffad453ff3f1-logs\") pod \"barbican-keystone-listener-7b4848f678-kfvpp\" (UID: \"6d8d8f21-1f1f-4431-8079-ffad453ff3f1\") " pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.936062 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-config\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.936093 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d8d8f21-1f1f-4431-8079-ffad453ff3f1-config-data\") pod \"barbican-keystone-listener-7b4848f678-kfvpp\" (UID: \"6d8d8f21-1f1f-4431-8079-ffad453ff3f1\") " pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.940948 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d8d8f21-1f1f-4431-8079-ffad453ff3f1-logs\") pod \"barbican-keystone-listener-7b4848f678-kfvpp\" (UID: \"6d8d8f21-1f1f-4431-8079-ffad453ff3f1\") " pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.945640 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6d8d8f21-1f1f-4431-8079-ffad453ff3f1-config-data-custom\") pod \"barbican-keystone-listener-7b4848f678-kfvpp\" (UID: \"6d8d8f21-1f1f-4431-8079-ffad453ff3f1\") " pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.951907 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d8d8f21-1f1f-4431-8079-ffad453ff3f1-config-data\") pod \"barbican-keystone-listener-7b4848f678-kfvpp\" (UID: \"6d8d8f21-1f1f-4431-8079-ffad453ff3f1\") " pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.954112 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-554df66f87-jscgb" Dec 03 06:46:56 crc kubenswrapper[4818]: I1203 06:46:56.970318 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d8d8f21-1f1f-4431-8079-ffad453ff3f1-combined-ca-bundle\") pod \"barbican-keystone-listener-7b4848f678-kfvpp\" (UID: \"6d8d8f21-1f1f-4431-8079-ffad453ff3f1\") " pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.032346 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpmwq\" (UniqueName: \"kubernetes.io/projected/6d8d8f21-1f1f-4431-8079-ffad453ff3f1-kube-api-access-kpmwq\") pod \"barbican-keystone-listener-7b4848f678-kfvpp\" (UID: \"6d8d8f21-1f1f-4431-8079-ffad453ff3f1\") " pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.038737 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.038802 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2r85\" (UniqueName: \"kubernetes.io/projected/add5819a-c589-4b78-9132-1a8166206a9e-kube-api-access-w2r85\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.038875 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-dns-svc\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.038898 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.038944 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-config\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.039012 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.039968 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.040530 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.041374 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-dns-svc\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.043994 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.044660 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-config\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.078222 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6ccc484b58-whz4r"] Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.079732 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.095745 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.107457 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.115903 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6ccc484b58-whz4r"] Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.118748 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2r85\" (UniqueName: \"kubernetes.io/projected/add5819a-c589-4b78-9132-1a8166206a9e-kube-api-access-w2r85\") pod \"dnsmasq-dns-85ff748b95-g9dwh\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.141754 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-combined-ca-bundle\") pod \"barbican-api-6ccc484b58-whz4r\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.141858 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-config-data\") pod \"barbican-api-6ccc484b58-whz4r\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.141876 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-config-data-custom\") pod \"barbican-api-6ccc484b58-whz4r\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.141942 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/947dba44-9121-4393-bad5-19439b447c99-logs\") pod \"barbican-api-6ccc484b58-whz4r\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.141973 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5fbp\" (UniqueName: \"kubernetes.io/projected/947dba44-9121-4393-bad5-19439b447c99-kube-api-access-k5fbp\") pod \"barbican-api-6ccc484b58-whz4r\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.243794 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-config-data\") pod \"barbican-api-6ccc484b58-whz4r\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.244249 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-config-data-custom\") pod \"barbican-api-6ccc484b58-whz4r\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.244382 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/947dba44-9121-4393-bad5-19439b447c99-logs\") pod \"barbican-api-6ccc484b58-whz4r\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.244440 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5fbp\" (UniqueName: \"kubernetes.io/projected/947dba44-9121-4393-bad5-19439b447c99-kube-api-access-k5fbp\") pod \"barbican-api-6ccc484b58-whz4r\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.244554 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-combined-ca-bundle\") pod \"barbican-api-6ccc484b58-whz4r\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.245533 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/947dba44-9121-4393-bad5-19439b447c99-logs\") pod \"barbican-api-6ccc484b58-whz4r\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.248764 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-combined-ca-bundle\") pod \"barbican-api-6ccc484b58-whz4r\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.249755 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-config-data-custom\") pod \"barbican-api-6ccc484b58-whz4r\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.249946 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-config-data\") pod \"barbican-api-6ccc484b58-whz4r\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.266610 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5fbp\" (UniqueName: \"kubernetes.io/projected/947dba44-9121-4393-bad5-19439b447c99-kube-api-access-k5fbp\") pod \"barbican-api-6ccc484b58-whz4r\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.415226 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:46:57 crc kubenswrapper[4818]: I1203 06:46:57.421499 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:46:58 crc kubenswrapper[4818]: I1203 06:46:58.450519 4818 generic.go:334] "Generic (PLEG): container finished" podID="20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3" containerID="2eec35e25d66cc807661730864b68b790a563c984227ff43593f4242c9b2ed83" exitCode=0 Dec 03 06:46:58 crc kubenswrapper[4818]: I1203 06:46:58.450573 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-978k5" event={"ID":"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3","Type":"ContainerDied","Data":"2eec35e25d66cc807661730864b68b790a563c984227ff43593f4242c9b2ed83"} Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.589631 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7454b897fd-qzp7b"] Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.591588 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.593689 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.595845 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.608205 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7454b897fd-qzp7b"] Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.688801 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e283b07-3849-4796-9b82-84e5ca0bf431-internal-tls-certs\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.688863 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e283b07-3849-4796-9b82-84e5ca0bf431-public-tls-certs\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.688897 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2z8t\" (UniqueName: \"kubernetes.io/projected/5e283b07-3849-4796-9b82-84e5ca0bf431-kube-api-access-s2z8t\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.688940 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e283b07-3849-4796-9b82-84e5ca0bf431-combined-ca-bundle\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.688958 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e283b07-3849-4796-9b82-84e5ca0bf431-config-data-custom\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.688981 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e283b07-3849-4796-9b82-84e5ca0bf431-config-data\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.688998 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e283b07-3849-4796-9b82-84e5ca0bf431-logs\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.796214 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e283b07-3849-4796-9b82-84e5ca0bf431-internal-tls-certs\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.796270 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e283b07-3849-4796-9b82-84e5ca0bf431-public-tls-certs\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.796312 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2z8t\" (UniqueName: \"kubernetes.io/projected/5e283b07-3849-4796-9b82-84e5ca0bf431-kube-api-access-s2z8t\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.796370 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e283b07-3849-4796-9b82-84e5ca0bf431-combined-ca-bundle\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.796395 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e283b07-3849-4796-9b82-84e5ca0bf431-config-data-custom\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.796427 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e283b07-3849-4796-9b82-84e5ca0bf431-config-data\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.796449 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e283b07-3849-4796-9b82-84e5ca0bf431-logs\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.797163 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e283b07-3849-4796-9b82-84e5ca0bf431-logs\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.805271 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e283b07-3849-4796-9b82-84e5ca0bf431-combined-ca-bundle\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.811460 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e283b07-3849-4796-9b82-84e5ca0bf431-config-data-custom\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.812151 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e283b07-3849-4796-9b82-84e5ca0bf431-public-tls-certs\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.819106 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e283b07-3849-4796-9b82-84e5ca0bf431-internal-tls-certs\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.833460 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2z8t\" (UniqueName: \"kubernetes.io/projected/5e283b07-3849-4796-9b82-84e5ca0bf431-kube-api-access-s2z8t\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.858579 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e283b07-3849-4796-9b82-84e5ca0bf431-config-data\") pod \"barbican-api-7454b897fd-qzp7b\" (UID: \"5e283b07-3849-4796-9b82-84e5ca0bf431\") " pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:46:59 crc kubenswrapper[4818]: I1203 06:46:59.915246 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:47:01 crc kubenswrapper[4818]: I1203 06:47:01.952197 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-978k5" Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.056918 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-db-sync-config-data\") pod \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.056972 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-combined-ca-bundle\") pod \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.057042 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-config-data\") pod \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.057089 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-etc-machine-id\") pod \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.057109 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-scripts\") pod \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.057159 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhh5p\" (UniqueName: \"kubernetes.io/projected/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-kube-api-access-xhh5p\") pod \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\" (UID: \"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3\") " Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.058017 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3" (UID: "20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.062965 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-scripts" (OuterVolumeSpecName: "scripts") pod "20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3" (UID: "20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.063012 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-kube-api-access-xhh5p" (OuterVolumeSpecName: "kube-api-access-xhh5p") pod "20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3" (UID: "20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3"). InnerVolumeSpecName "kube-api-access-xhh5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.063503 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3" (UID: "20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.092409 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3" (UID: "20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.110789 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-config-data" (OuterVolumeSpecName: "config-data") pod "20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3" (UID: "20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.160253 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhh5p\" (UniqueName: \"kubernetes.io/projected/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-kube-api-access-xhh5p\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.160312 4818 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.160325 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.160338 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.160348 4818 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.160360 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.525136 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-978k5" event={"ID":"20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3","Type":"ContainerDied","Data":"7bb6cbbd03d8420de615001cd19501e708e819de1225b337dc628b6666a4e468"} Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.525176 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bb6cbbd03d8420de615001cd19501e708e819de1225b337dc628b6666a4e468" Dec 03 06:47:02 crc kubenswrapper[4818]: I1203 06:47:02.525190 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-978k5" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.447905 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 06:47:03 crc kubenswrapper[4818]: E1203 06:47:03.449574 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3" containerName="cinder-db-sync" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.449596 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3" containerName="cinder-db-sync" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.450020 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3" containerName="cinder-db-sync" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.481042 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.485632 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-m2q9c" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.487641 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.489187 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.518996 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.571891 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.572780 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.572997 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-config-data\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.573181 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf9kj\" (UniqueName: \"kubernetes.io/projected/952372cb-1c5f-4aa9-a751-3526b5ced9a7-kube-api-access-vf9kj\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.577532 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/952372cb-1c5f-4aa9-a751-3526b5ced9a7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.577694 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-scripts\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.579519 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.641141 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-g9dwh"] Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.653074 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-qzppx"] Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.654849 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.681288 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.681344 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/952372cb-1c5f-4aa9-a751-3526b5ced9a7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.681366 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-scripts\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.681408 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.681432 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.681457 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-config\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.681476 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.681512 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-config-data\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.681540 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf9kj\" (UniqueName: \"kubernetes.io/projected/952372cb-1c5f-4aa9-a751-3526b5ced9a7-kube-api-access-vf9kj\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.681585 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.681603 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnckx\" (UniqueName: \"kubernetes.io/projected/c27270e3-defd-4d5b-abcc-2f717189dbd5-kube-api-access-hnckx\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.681621 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.681726 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/952372cb-1c5f-4aa9-a751-3526b5ced9a7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.686975 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-qzppx"] Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.692795 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.701544 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-scripts\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.702093 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.713837 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-config-data\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.757180 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf9kj\" (UniqueName: \"kubernetes.io/projected/952372cb-1c5f-4aa9-a751-3526b5ced9a7-kube-api-access-vf9kj\") pod \"cinder-scheduler-0\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.800304 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.800337 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnckx\" (UniqueName: \"kubernetes.io/projected/c27270e3-defd-4d5b-abcc-2f717189dbd5-kube-api-access-hnckx\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.800357 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.800395 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.800451 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.800474 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-config\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.801418 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-config\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.802673 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.803507 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.807800 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.813080 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.851576 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.853110 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.869838 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.876215 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnckx\" (UniqueName: \"kubernetes.io/projected/c27270e3-defd-4d5b-abcc-2f717189dbd5-kube-api-access-hnckx\") pod \"dnsmasq-dns-5c9776ccc5-qzppx\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.879984 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.902040 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wnrh\" (UniqueName: \"kubernetes.io/projected/fc1579d3-d123-45ce-9b92-9486a73de5db-kube-api-access-4wnrh\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.902084 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc1579d3-d123-45ce-9b92-9486a73de5db-logs\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.902112 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-config-data-custom\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.902132 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-config-data\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.902160 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-scripts\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.902175 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.902196 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc1579d3-d123-45ce-9b92-9486a73de5db-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:03 crc kubenswrapper[4818]: I1203 06:47:03.912069 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 06:47:04 crc kubenswrapper[4818]: I1203 06:47:04.003517 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:04 crc kubenswrapper[4818]: I1203 06:47:04.004263 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wnrh\" (UniqueName: \"kubernetes.io/projected/fc1579d3-d123-45ce-9b92-9486a73de5db-kube-api-access-4wnrh\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:04 crc kubenswrapper[4818]: I1203 06:47:04.004324 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc1579d3-d123-45ce-9b92-9486a73de5db-logs\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:04 crc kubenswrapper[4818]: I1203 06:47:04.004361 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-config-data-custom\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:04 crc kubenswrapper[4818]: I1203 06:47:04.004391 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-config-data\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:04 crc kubenswrapper[4818]: I1203 06:47:04.004432 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-scripts\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:04 crc kubenswrapper[4818]: I1203 06:47:04.004455 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:04 crc kubenswrapper[4818]: I1203 06:47:04.004482 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc1579d3-d123-45ce-9b92-9486a73de5db-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:04 crc kubenswrapper[4818]: I1203 06:47:04.004675 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc1579d3-d123-45ce-9b92-9486a73de5db-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:04 crc kubenswrapper[4818]: I1203 06:47:04.005499 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc1579d3-d123-45ce-9b92-9486a73de5db-logs\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:04 crc kubenswrapper[4818]: I1203 06:47:04.008500 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-config-data-custom\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:04 crc kubenswrapper[4818]: I1203 06:47:04.011562 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:04 crc kubenswrapper[4818]: I1203 06:47:04.016695 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-scripts\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:04 crc kubenswrapper[4818]: I1203 06:47:04.018067 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-config-data\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:04 crc kubenswrapper[4818]: I1203 06:47:04.026649 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wnrh\" (UniqueName: \"kubernetes.io/projected/fc1579d3-d123-45ce-9b92-9486a73de5db-kube-api-access-4wnrh\") pod \"cinder-api-0\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " pod="openstack/cinder-api-0" Dec 03 06:47:04 crc kubenswrapper[4818]: I1203 06:47:04.217131 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 06:47:05 crc kubenswrapper[4818]: I1203 06:47:05.395926 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 06:47:05 crc kubenswrapper[4818]: E1203 06:47:05.676284 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="575fe180-2a23-48e9-b5cf-b90e83528e5e" Dec 03 06:47:05 crc kubenswrapper[4818]: I1203 06:47:05.975281 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-554df66f87-jscgb"] Dec 03 06:47:05 crc kubenswrapper[4818]: W1203 06:47:05.989920 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod947dba44_9121_4393_bad5_19439b447c99.slice/crio-cedf1752781a1598e7d334193cd4a7af0413eee158f699001ae583daa196b62a WatchSource:0}: Error finding container cedf1752781a1598e7d334193cd4a7af0413eee158f699001ae583daa196b62a: Status 404 returned error can't find the container with id cedf1752781a1598e7d334193cd4a7af0413eee158f699001ae583daa196b62a Dec 03 06:47:05 crc kubenswrapper[4818]: I1203 06:47:05.992779 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7b4848f678-kfvpp"] Dec 03 06:47:06 crc kubenswrapper[4818]: W1203 06:47:06.043545 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8e58804_c7d8_48ce_9db3_6c3ab56f1c13.slice/crio-3257a2a3d20fc1802559a7ec0ca19402e1f1112be2862b7204c907b280b84053 WatchSource:0}: Error finding container 3257a2a3d20fc1802559a7ec0ca19402e1f1112be2862b7204c907b280b84053: Status 404 returned error can't find the container with id 3257a2a3d20fc1802559a7ec0ca19402e1f1112be2862b7204c907b280b84053 Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.068302 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6ccc484b58-whz4r"] Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.401048 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.411633 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7454b897fd-qzp7b"] Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.444546 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.466593 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-qzppx"] Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.525221 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-g9dwh"] Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.709025 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fc1579d3-d123-45ce-9b92-9486a73de5db","Type":"ContainerStarted","Data":"521ce8c5c3ee54dfdf2c2d3aad5f0f037b014f298eef6513e9ca6f2082385b46"} Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.736092 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"952372cb-1c5f-4aa9-a751-3526b5ced9a7","Type":"ContainerStarted","Data":"679ef967b9a2abd7f1d98925c44b174d94fbc61e8de7dc02988241b165c24c3c"} Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.792653 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" event={"ID":"add5819a-c589-4b78-9132-1a8166206a9e","Type":"ContainerStarted","Data":"92a88ccf335af400d4ce028004a469875a96c8001fc65e0cfbd2b1e1c7f2e232"} Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.845052 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" event={"ID":"6d8d8f21-1f1f-4431-8079-ffad453ff3f1","Type":"ContainerStarted","Data":"f0588ffaf394db747da6f58351189e47704bf39967f51a77ad0cd44ec6f11a99"} Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.877009 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7454b897fd-qzp7b" event={"ID":"5e283b07-3849-4796-9b82-84e5ca0bf431","Type":"ContainerStarted","Data":"e44dacb61953c76d4e57d4d4f212551e2bb38ddb152fbb86366e3974f4a3cbd4"} Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.880504 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-554df66f87-jscgb" event={"ID":"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13","Type":"ContainerStarted","Data":"3257a2a3d20fc1802559a7ec0ca19402e1f1112be2862b7204c907b280b84053"} Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.914965 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"575fe180-2a23-48e9-b5cf-b90e83528e5e","Type":"ContainerStarted","Data":"476e7a3873cbe0683086d8ce5ed7237071f9030a7eedc3126f748c7e3af83718"} Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.915122 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="575fe180-2a23-48e9-b5cf-b90e83528e5e" containerName="ceilometer-notification-agent" containerID="cri-o://8f699e105a92636cbeab2633821e2e90792849bdc5323a39d6e9a3ee12f8fa9c" gracePeriod=30 Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.915349 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.915615 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="575fe180-2a23-48e9-b5cf-b90e83528e5e" containerName="proxy-httpd" containerID="cri-o://476e7a3873cbe0683086d8ce5ed7237071f9030a7eedc3126f748c7e3af83718" gracePeriod=30 Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.915659 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="575fe180-2a23-48e9-b5cf-b90e83528e5e" containerName="sg-core" containerID="cri-o://693c121be8011128371d293e5fc85539a1e5af22221f56702ec76f8eb5db0714" gracePeriod=30 Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.929706 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6ccc484b58-whz4r" event={"ID":"947dba44-9121-4393-bad5-19439b447c99","Type":"ContainerStarted","Data":"45399f78c19dec4c370fdf201666cc6eca7f195c8cf3e9c3ec077e9de0263c3d"} Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.929751 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6ccc484b58-whz4r" event={"ID":"947dba44-9121-4393-bad5-19439b447c99","Type":"ContainerStarted","Data":"cedf1752781a1598e7d334193cd4a7af0413eee158f699001ae583daa196b62a"} Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.932411 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" event={"ID":"c27270e3-defd-4d5b-abcc-2f717189dbd5","Type":"ContainerStarted","Data":"5afd38016a169387ac6fa5f4edab9bc61d23d03ea0e15433dd46acac97b5a6e5"} Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.940242 4818 generic.go:334] "Generic (PLEG): container finished" podID="9d0cba79-7a13-40ee-ab38-ae093b204a9c" containerID="703567b992f5fda96a6c66e4c3c5c7af52da72a3847c84b247816dc6adfdcd7a" exitCode=137 Dec 03 06:47:06 crc kubenswrapper[4818]: I1203 06:47:06.940306 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fbcd7d5fc-6stq9" event={"ID":"9d0cba79-7a13-40ee-ab38-ae093b204a9c","Type":"ContainerDied","Data":"703567b992f5fda96a6c66e4c3c5c7af52da72a3847c84b247816dc6adfdcd7a"} Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.523026 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.664798 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.764673 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f18580e-5450-4edb-b266-734372950182-scripts\") pod \"2f18580e-5450-4edb-b266-734372950182\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.764722 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f18580e-5450-4edb-b266-734372950182-config-data\") pod \"2f18580e-5450-4edb-b266-734372950182\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.764783 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f18580e-5450-4edb-b266-734372950182-logs\") pod \"2f18580e-5450-4edb-b266-734372950182\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.764807 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2f18580e-5450-4edb-b266-734372950182-horizon-secret-key\") pod \"2f18580e-5450-4edb-b266-734372950182\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.764995 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmbqm\" (UniqueName: \"kubernetes.io/projected/2f18580e-5450-4edb-b266-734372950182-kube-api-access-vmbqm\") pod \"2f18580e-5450-4edb-b266-734372950182\" (UID: \"2f18580e-5450-4edb-b266-734372950182\") " Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.767619 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f18580e-5450-4edb-b266-734372950182-logs" (OuterVolumeSpecName: "logs") pod "2f18580e-5450-4edb-b266-734372950182" (UID: "2f18580e-5450-4edb-b266-734372950182"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.793118 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f18580e-5450-4edb-b266-734372950182-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "2f18580e-5450-4edb-b266-734372950182" (UID: "2f18580e-5450-4edb-b266-734372950182"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.793131 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f18580e-5450-4edb-b266-734372950182-kube-api-access-vmbqm" (OuterVolumeSpecName: "kube-api-access-vmbqm") pod "2f18580e-5450-4edb-b266-734372950182" (UID: "2f18580e-5450-4edb-b266-734372950182"). InnerVolumeSpecName "kube-api-access-vmbqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.814988 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f18580e-5450-4edb-b266-734372950182-scripts" (OuterVolumeSpecName: "scripts") pod "2f18580e-5450-4edb-b266-734372950182" (UID: "2f18580e-5450-4edb-b266-734372950182"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.838946 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.865465 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f18580e-5450-4edb-b266-734372950182-config-data" (OuterVolumeSpecName: "config-data") pod "2f18580e-5450-4edb-b266-734372950182" (UID: "2f18580e-5450-4edb-b266-734372950182"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.867087 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f18580e-5450-4edb-b266-734372950182-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.867106 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f18580e-5450-4edb-b266-734372950182-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.867117 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f18580e-5450-4edb-b266-734372950182-logs\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.867125 4818 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2f18580e-5450-4edb-b266-734372950182-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.867133 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmbqm\" (UniqueName: \"kubernetes.io/projected/2f18580e-5450-4edb-b266-734372950182-kube-api-access-vmbqm\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.972554 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9d0cba79-7a13-40ee-ab38-ae093b204a9c-horizon-secret-key\") pod \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.972643 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xp7qh\" (UniqueName: \"kubernetes.io/projected/9d0cba79-7a13-40ee-ab38-ae093b204a9c-kube-api-access-xp7qh\") pod \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.972675 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d0cba79-7a13-40ee-ab38-ae093b204a9c-scripts\") pod \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.972757 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d0cba79-7a13-40ee-ab38-ae093b204a9c-config-data\") pod \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.972842 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d0cba79-7a13-40ee-ab38-ae093b204a9c-logs\") pod \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\" (UID: \"9d0cba79-7a13-40ee-ab38-ae093b204a9c\") " Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.973586 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d0cba79-7a13-40ee-ab38-ae093b204a9c-logs" (OuterVolumeSpecName: "logs") pod "9d0cba79-7a13-40ee-ab38-ae093b204a9c" (UID: "9d0cba79-7a13-40ee-ab38-ae093b204a9c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.980071 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d0cba79-7a13-40ee-ab38-ae093b204a9c-kube-api-access-xp7qh" (OuterVolumeSpecName: "kube-api-access-xp7qh") pod "9d0cba79-7a13-40ee-ab38-ae093b204a9c" (UID: "9d0cba79-7a13-40ee-ab38-ae093b204a9c"). InnerVolumeSpecName "kube-api-access-xp7qh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:07 crc kubenswrapper[4818]: I1203 06:47:07.985980 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d0cba79-7a13-40ee-ab38-ae093b204a9c-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "9d0cba79-7a13-40ee-ab38-ae093b204a9c" (UID: "9d0cba79-7a13-40ee-ab38-ae093b204a9c"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.008984 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d0cba79-7a13-40ee-ab38-ae093b204a9c-scripts" (OuterVolumeSpecName: "scripts") pod "9d0cba79-7a13-40ee-ab38-ae093b204a9c" (UID: "9d0cba79-7a13-40ee-ab38-ae093b204a9c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.009223 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d0cba79-7a13-40ee-ab38-ae093b204a9c-config-data" (OuterVolumeSpecName: "config-data") pod "9d0cba79-7a13-40ee-ab38-ae093b204a9c" (UID: "9d0cba79-7a13-40ee-ab38-ae093b204a9c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.030402 4818 generic.go:334] "Generic (PLEG): container finished" podID="575fe180-2a23-48e9-b5cf-b90e83528e5e" containerID="476e7a3873cbe0683086d8ce5ed7237071f9030a7eedc3126f748c7e3af83718" exitCode=0 Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.030441 4818 generic.go:334] "Generic (PLEG): container finished" podID="575fe180-2a23-48e9-b5cf-b90e83528e5e" containerID="693c121be8011128371d293e5fc85539a1e5af22221f56702ec76f8eb5db0714" exitCode=2 Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.030508 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"575fe180-2a23-48e9-b5cf-b90e83528e5e","Type":"ContainerDied","Data":"476e7a3873cbe0683086d8ce5ed7237071f9030a7eedc3126f748c7e3af83718"} Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.030537 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"575fe180-2a23-48e9-b5cf-b90e83528e5e","Type":"ContainerDied","Data":"693c121be8011128371d293e5fc85539a1e5af22221f56702ec76f8eb5db0714"} Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.033092 4818 generic.go:334] "Generic (PLEG): container finished" podID="2f18580e-5450-4edb-b266-734372950182" containerID="707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115" exitCode=137 Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.033123 4818 generic.go:334] "Generic (PLEG): container finished" podID="2f18580e-5450-4edb-b266-734372950182" containerID="9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc" exitCode=137 Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.033163 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-797779bc9c-df4k9" event={"ID":"2f18580e-5450-4edb-b266-734372950182","Type":"ContainerDied","Data":"707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115"} Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.033188 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-797779bc9c-df4k9" event={"ID":"2f18580e-5450-4edb-b266-734372950182","Type":"ContainerDied","Data":"9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc"} Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.033200 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-797779bc9c-df4k9" event={"ID":"2f18580e-5450-4edb-b266-734372950182","Type":"ContainerDied","Data":"4de6cea32430cd98c694d5818917fcda1879030caecf10d9124e3eb917a1a81b"} Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.033218 4818 scope.go:117] "RemoveContainer" containerID="707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.033372 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-797779bc9c-df4k9" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.048164 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6ccc484b58-whz4r" event={"ID":"947dba44-9121-4393-bad5-19439b447c99","Type":"ContainerStarted","Data":"77d24734d359fc520b481b5a57b8bcea5c85b8033ddf54f54f4d688f22546aa9"} Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.050860 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.060330 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.074927 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d0cba79-7a13-40ee-ab38-ae093b204a9c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.074962 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d0cba79-7a13-40ee-ab38-ae093b204a9c-logs\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.074974 4818 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9d0cba79-7a13-40ee-ab38-ae093b204a9c-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.074986 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xp7qh\" (UniqueName: \"kubernetes.io/projected/9d0cba79-7a13-40ee-ab38-ae093b204a9c-kube-api-access-xp7qh\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.074997 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d0cba79-7a13-40ee-ab38-ae093b204a9c-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.078595 4818 generic.go:334] "Generic (PLEG): container finished" podID="c27270e3-defd-4d5b-abcc-2f717189dbd5" containerID="29f6bb64fef338a65b9e3e325130da2d0822d016e8aaa5bd246a9dcaf965d0b7" exitCode=0 Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.079007 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" event={"ID":"c27270e3-defd-4d5b-abcc-2f717189dbd5","Type":"ContainerDied","Data":"29f6bb64fef338a65b9e3e325130da2d0822d016e8aaa5bd246a9dcaf965d0b7"} Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.079474 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6ccc484b58-whz4r" podStartSLOduration=11.079461004 podStartE2EDuration="11.079461004s" podCreationTimestamp="2025-12-03 06:46:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:47:08.078527401 +0000 UTC m=+1185.770136153" watchObservedRunningTime="2025-12-03 06:47:08.079461004 +0000 UTC m=+1185.771069756" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.118554 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7454b897fd-qzp7b" event={"ID":"5e283b07-3849-4796-9b82-84e5ca0bf431","Type":"ContainerStarted","Data":"d7669a00fa3660db0c9173040f37de51e0f4a2b7c94a402add9283d0533b4eaa"} Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.118611 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7454b897fd-qzp7b" event={"ID":"5e283b07-3849-4796-9b82-84e5ca0bf431","Type":"ContainerStarted","Data":"129863e5449f3f47b257db236c094cf01bfdffac9b0f31394f8efa9a934ec6ea"} Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.118651 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.118671 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.145085 4818 generic.go:334] "Generic (PLEG): container finished" podID="9d0cba79-7a13-40ee-ab38-ae093b204a9c" containerID="42d52b447fd707dfb9a71ccd0ff746cf0e8dda0c5ba69b869f28d7f7cd076848" exitCode=137 Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.145175 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fbcd7d5fc-6stq9" event={"ID":"9d0cba79-7a13-40ee-ab38-ae093b204a9c","Type":"ContainerDied","Data":"42d52b447fd707dfb9a71ccd0ff746cf0e8dda0c5ba69b869f28d7f7cd076848"} Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.145202 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fbcd7d5fc-6stq9" event={"ID":"9d0cba79-7a13-40ee-ab38-ae093b204a9c","Type":"ContainerDied","Data":"43fb75ae5439022048bfc06ddf6cb52024a6a7f9401d56d1da180212077cfd88"} Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.145275 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fbcd7d5fc-6stq9" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.149030 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7454b897fd-qzp7b" podStartSLOduration=9.1490148 podStartE2EDuration="9.1490148s" podCreationTimestamp="2025-12-03 06:46:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:47:08.135851657 +0000 UTC m=+1185.827460419" watchObservedRunningTime="2025-12-03 06:47:08.1490148 +0000 UTC m=+1185.840623552" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.151474 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fc1579d3-d123-45ce-9b92-9486a73de5db","Type":"ContainerStarted","Data":"1daa47ee7b9c9b5b65ffdd8ef03140fe7f08328ddff57e821221885062ad32bb"} Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.158981 4818 generic.go:334] "Generic (PLEG): container finished" podID="add5819a-c589-4b78-9132-1a8166206a9e" containerID="27ea0b0ba21e649a59e063dc1ddb35f4764243d1858cd159d67d72243d4fdb83" exitCode=0 Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.159047 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" event={"ID":"add5819a-c589-4b78-9132-1a8166206a9e","Type":"ContainerDied","Data":"27ea0b0ba21e649a59e063dc1ddb35f4764243d1858cd159d67d72243d4fdb83"} Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.272606 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-797779bc9c-df4k9"] Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.286393 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-797779bc9c-df4k9"] Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.337119 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7fbcd7d5fc-6stq9"] Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.345907 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7fbcd7d5fc-6stq9"] Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.509837 4818 scope.go:117] "RemoveContainer" containerID="9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.553441 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.577528 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.750974 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f18580e-5450-4edb-b266-734372950182" path="/var/lib/kubelet/pods/2f18580e-5450-4edb-b266-734372950182/volumes" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.751682 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d0cba79-7a13-40ee-ab38-ae093b204a9c" path="/var/lib/kubelet/pods/9d0cba79-7a13-40ee-ab38-ae093b204a9c/volumes" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.939447 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.950943 4818 scope.go:117] "RemoveContainer" containerID="707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115" Dec 03 06:47:08 crc kubenswrapper[4818]: E1203 06:47:08.956859 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115\": container with ID starting with 707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115 not found: ID does not exist" containerID="707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.956894 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115"} err="failed to get container status \"707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115\": rpc error: code = NotFound desc = could not find container \"707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115\": container with ID starting with 707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115 not found: ID does not exist" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.956917 4818 scope.go:117] "RemoveContainer" containerID="9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc" Dec 03 06:47:08 crc kubenswrapper[4818]: E1203 06:47:08.959135 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc\": container with ID starting with 9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc not found: ID does not exist" containerID="9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.959161 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc"} err="failed to get container status \"9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc\": rpc error: code = NotFound desc = could not find container \"9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc\": container with ID starting with 9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc not found: ID does not exist" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.959176 4818 scope.go:117] "RemoveContainer" containerID="707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.961808 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115"} err="failed to get container status \"707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115\": rpc error: code = NotFound desc = could not find container \"707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115\": container with ID starting with 707f1068d5132be69685d981b3e5cecc4759dba8bfc3679206afb0f5359af115 not found: ID does not exist" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.961841 4818 scope.go:117] "RemoveContainer" containerID="9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.962267 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc"} err="failed to get container status \"9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc\": rpc error: code = NotFound desc = could not find container \"9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc\": container with ID starting with 9fc50501c5f5e5b14b3e21cfc7518e99f797e10da4c1ab8fc6536770785631fc not found: ID does not exist" Dec 03 06:47:08 crc kubenswrapper[4818]: I1203 06:47:08.962286 4818 scope.go:117] "RemoveContainer" containerID="42d52b447fd707dfb9a71ccd0ff746cf0e8dda0c5ba69b869f28d7f7cd076848" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.097962 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-ovsdbserver-sb\") pod \"add5819a-c589-4b78-9132-1a8166206a9e\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.098129 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-dns-swift-storage-0\") pod \"add5819a-c589-4b78-9132-1a8166206a9e\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.098265 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-config\") pod \"add5819a-c589-4b78-9132-1a8166206a9e\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.098316 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-ovsdbserver-nb\") pod \"add5819a-c589-4b78-9132-1a8166206a9e\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.098409 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2r85\" (UniqueName: \"kubernetes.io/projected/add5819a-c589-4b78-9132-1a8166206a9e-kube-api-access-w2r85\") pod \"add5819a-c589-4b78-9132-1a8166206a9e\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.098530 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-dns-svc\") pod \"add5819a-c589-4b78-9132-1a8166206a9e\" (UID: \"add5819a-c589-4b78-9132-1a8166206a9e\") " Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.107048 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/add5819a-c589-4b78-9132-1a8166206a9e-kube-api-access-w2r85" (OuterVolumeSpecName: "kube-api-access-w2r85") pod "add5819a-c589-4b78-9132-1a8166206a9e" (UID: "add5819a-c589-4b78-9132-1a8166206a9e"). InnerVolumeSpecName "kube-api-access-w2r85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.127043 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-config" (OuterVolumeSpecName: "config") pod "add5819a-c589-4b78-9132-1a8166206a9e" (UID: "add5819a-c589-4b78-9132-1a8166206a9e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.143399 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "add5819a-c589-4b78-9132-1a8166206a9e" (UID: "add5819a-c589-4b78-9132-1a8166206a9e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.149213 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "add5819a-c589-4b78-9132-1a8166206a9e" (UID: "add5819a-c589-4b78-9132-1a8166206a9e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.155603 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "add5819a-c589-4b78-9132-1a8166206a9e" (UID: "add5819a-c589-4b78-9132-1a8166206a9e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.171586 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"952372cb-1c5f-4aa9-a751-3526b5ced9a7","Type":"ContainerStarted","Data":"080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3"} Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.173022 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" event={"ID":"add5819a-c589-4b78-9132-1a8166206a9e","Type":"ContainerDied","Data":"92a88ccf335af400d4ce028004a469875a96c8001fc65e0cfbd2b1e1c7f2e232"} Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.173093 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-g9dwh" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.179501 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" event={"ID":"c27270e3-defd-4d5b-abcc-2f717189dbd5","Type":"ContainerStarted","Data":"963ea9b3f0e8ad84c7d038bccd3864c2c651117a41c9203d8cbbc239c00df7fe"} Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.182472 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.186887 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fc1579d3-d123-45ce-9b92-9486a73de5db","Type":"ContainerStarted","Data":"2d744852d7d9feafa9035e2bd6a63c720308ca43860922303d7d441e7cd437de"} Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.187178 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="fc1579d3-d123-45ce-9b92-9486a73de5db" containerName="cinder-api-log" containerID="cri-o://1daa47ee7b9c9b5b65ffdd8ef03140fe7f08328ddff57e821221885062ad32bb" gracePeriod=30 Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.187263 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.187297 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="fc1579d3-d123-45ce-9b92-9486a73de5db" containerName="cinder-api" containerID="cri-o://2d744852d7d9feafa9035e2bd6a63c720308ca43860922303d7d441e7cd437de" gracePeriod=30 Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.196769 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "add5819a-c589-4b78-9132-1a8166206a9e" (UID: "add5819a-c589-4b78-9132-1a8166206a9e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.198352 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" podStartSLOduration=6.198338713 podStartE2EDuration="6.198338713s" podCreationTimestamp="2025-12-03 06:47:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:47:09.1953392 +0000 UTC m=+1186.886947952" watchObservedRunningTime="2025-12-03 06:47:09.198338713 +0000 UTC m=+1186.889947465" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.204890 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.204918 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.204930 4818 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.204941 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.204949 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/add5819a-c589-4b78-9132-1a8166206a9e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.204958 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2r85\" (UniqueName: \"kubernetes.io/projected/add5819a-c589-4b78-9132-1a8166206a9e-kube-api-access-w2r85\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.221707 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.221692017 podStartE2EDuration="6.221692017s" podCreationTimestamp="2025-12-03 06:47:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:47:09.219121223 +0000 UTC m=+1186.910729985" watchObservedRunningTime="2025-12-03 06:47:09.221692017 +0000 UTC m=+1186.913300769" Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.541942 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-g9dwh"] Dec 03 06:47:09 crc kubenswrapper[4818]: I1203 06:47:09.550555 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-g9dwh"] Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.046151 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-77cd7f4557-lnszf" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.154919 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-69b59f8dd-sckpz"] Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.155233 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-69b59f8dd-sckpz" podUID="ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b" containerName="neutron-api" containerID="cri-o://43fce69ae082d44a241742d1be5d81cb21889f23487c23b323a4b77d499b7453" gracePeriod=30 Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.156235 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-69b59f8dd-sckpz" podUID="ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b" containerName="neutron-httpd" containerID="cri-o://45c9fba1cd213a75cebfe1bfc9c9756645c2b5a012d0442c0ffee17a3bc40bab" gracePeriod=30 Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.174998 4818 scope.go:117] "RemoveContainer" containerID="703567b992f5fda96a6c66e4c3c5c7af52da72a3847c84b247816dc6adfdcd7a" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.266067 4818 generic.go:334] "Generic (PLEG): container finished" podID="fc1579d3-d123-45ce-9b92-9486a73de5db" containerID="2d744852d7d9feafa9035e2bd6a63c720308ca43860922303d7d441e7cd437de" exitCode=0 Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.266098 4818 generic.go:334] "Generic (PLEG): container finished" podID="fc1579d3-d123-45ce-9b92-9486a73de5db" containerID="1daa47ee7b9c9b5b65ffdd8ef03140fe7f08328ddff57e821221885062ad32bb" exitCode=143 Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.267055 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fc1579d3-d123-45ce-9b92-9486a73de5db","Type":"ContainerDied","Data":"2d744852d7d9feafa9035e2bd6a63c720308ca43860922303d7d441e7cd437de"} Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.267097 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fc1579d3-d123-45ce-9b92-9486a73de5db","Type":"ContainerDied","Data":"1daa47ee7b9c9b5b65ffdd8ef03140fe7f08328ddff57e821221885062ad32bb"} Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.317643 4818 scope.go:117] "RemoveContainer" containerID="42d52b447fd707dfb9a71ccd0ff746cf0e8dda0c5ba69b869f28d7f7cd076848" Dec 03 06:47:10 crc kubenswrapper[4818]: E1203 06:47:10.324944 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42d52b447fd707dfb9a71ccd0ff746cf0e8dda0c5ba69b869f28d7f7cd076848\": container with ID starting with 42d52b447fd707dfb9a71ccd0ff746cf0e8dda0c5ba69b869f28d7f7cd076848 not found: ID does not exist" containerID="42d52b447fd707dfb9a71ccd0ff746cf0e8dda0c5ba69b869f28d7f7cd076848" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.325035 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42d52b447fd707dfb9a71ccd0ff746cf0e8dda0c5ba69b869f28d7f7cd076848"} err="failed to get container status \"42d52b447fd707dfb9a71ccd0ff746cf0e8dda0c5ba69b869f28d7f7cd076848\": rpc error: code = NotFound desc = could not find container \"42d52b447fd707dfb9a71ccd0ff746cf0e8dda0c5ba69b869f28d7f7cd076848\": container with ID starting with 42d52b447fd707dfb9a71ccd0ff746cf0e8dda0c5ba69b869f28d7f7cd076848 not found: ID does not exist" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.325096 4818 scope.go:117] "RemoveContainer" containerID="703567b992f5fda96a6c66e4c3c5c7af52da72a3847c84b247816dc6adfdcd7a" Dec 03 06:47:10 crc kubenswrapper[4818]: E1203 06:47:10.326167 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"703567b992f5fda96a6c66e4c3c5c7af52da72a3847c84b247816dc6adfdcd7a\": container with ID starting with 703567b992f5fda96a6c66e4c3c5c7af52da72a3847c84b247816dc6adfdcd7a not found: ID does not exist" containerID="703567b992f5fda96a6c66e4c3c5c7af52da72a3847c84b247816dc6adfdcd7a" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.326198 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"703567b992f5fda96a6c66e4c3c5c7af52da72a3847c84b247816dc6adfdcd7a"} err="failed to get container status \"703567b992f5fda96a6c66e4c3c5c7af52da72a3847c84b247816dc6adfdcd7a\": rpc error: code = NotFound desc = could not find container \"703567b992f5fda96a6c66e4c3c5c7af52da72a3847c84b247816dc6adfdcd7a\": container with ID starting with 703567b992f5fda96a6c66e4c3c5c7af52da72a3847c84b247816dc6adfdcd7a not found: ID does not exist" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.326216 4818 scope.go:117] "RemoveContainer" containerID="27ea0b0ba21e649a59e063dc1ddb35f4764243d1858cd159d67d72243d4fdb83" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.481916 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.656798 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-scripts\") pod \"fc1579d3-d123-45ce-9b92-9486a73de5db\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.656925 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-config-data\") pod \"fc1579d3-d123-45ce-9b92-9486a73de5db\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.656950 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc1579d3-d123-45ce-9b92-9486a73de5db-etc-machine-id\") pod \"fc1579d3-d123-45ce-9b92-9486a73de5db\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.656972 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wnrh\" (UniqueName: \"kubernetes.io/projected/fc1579d3-d123-45ce-9b92-9486a73de5db-kube-api-access-4wnrh\") pod \"fc1579d3-d123-45ce-9b92-9486a73de5db\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.656993 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-combined-ca-bundle\") pod \"fc1579d3-d123-45ce-9b92-9486a73de5db\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.657114 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc1579d3-d123-45ce-9b92-9486a73de5db-logs\") pod \"fc1579d3-d123-45ce-9b92-9486a73de5db\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.657168 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-config-data-custom\") pod \"fc1579d3-d123-45ce-9b92-9486a73de5db\" (UID: \"fc1579d3-d123-45ce-9b92-9486a73de5db\") " Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.659727 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc1579d3-d123-45ce-9b92-9486a73de5db-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "fc1579d3-d123-45ce-9b92-9486a73de5db" (UID: "fc1579d3-d123-45ce-9b92-9486a73de5db"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.662063 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc1579d3-d123-45ce-9b92-9486a73de5db-logs" (OuterVolumeSpecName: "logs") pod "fc1579d3-d123-45ce-9b92-9486a73de5db" (UID: "fc1579d3-d123-45ce-9b92-9486a73de5db"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.664133 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fc1579d3-d123-45ce-9b92-9486a73de5db" (UID: "fc1579d3-d123-45ce-9b92-9486a73de5db"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.664370 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-scripts" (OuterVolumeSpecName: "scripts") pod "fc1579d3-d123-45ce-9b92-9486a73de5db" (UID: "fc1579d3-d123-45ce-9b92-9486a73de5db"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.673339 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc1579d3-d123-45ce-9b92-9486a73de5db-kube-api-access-4wnrh" (OuterVolumeSpecName: "kube-api-access-4wnrh") pod "fc1579d3-d123-45ce-9b92-9486a73de5db" (UID: "fc1579d3-d123-45ce-9b92-9486a73de5db"). InnerVolumeSpecName "kube-api-access-4wnrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.717148 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc1579d3-d123-45ce-9b92-9486a73de5db" (UID: "fc1579d3-d123-45ce-9b92-9486a73de5db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.759021 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc1579d3-d123-45ce-9b92-9486a73de5db-logs\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.759062 4818 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.759078 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.759091 4818 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc1579d3-d123-45ce-9b92-9486a73de5db-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.759103 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wnrh\" (UniqueName: \"kubernetes.io/projected/fc1579d3-d123-45ce-9b92-9486a73de5db-kube-api-access-4wnrh\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.759119 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.761094 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="add5819a-c589-4b78-9132-1a8166206a9e" path="/var/lib/kubelet/pods/add5819a-c589-4b78-9132-1a8166206a9e/volumes" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.796937 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-config-data" (OuterVolumeSpecName: "config-data") pod "fc1579d3-d123-45ce-9b92-9486a73de5db" (UID: "fc1579d3-d123-45ce-9b92-9486a73de5db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.946806 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc1579d3-d123-45ce-9b92-9486a73de5db-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.971254 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-68b68c8c84-qt25s" Dec 03 06:47:10 crc kubenswrapper[4818]: I1203 06:47:10.975040 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.050351 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-574dc6486d-nvkzv"] Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.283307 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fc1579d3-d123-45ce-9b92-9486a73de5db","Type":"ContainerDied","Data":"521ce8c5c3ee54dfdf2c2d3aad5f0f037b014f298eef6513e9ca6f2082385b46"} Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.283365 4818 scope.go:117] "RemoveContainer" containerID="2d744852d7d9feafa9035e2bd6a63c720308ca43860922303d7d441e7cd437de" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.283380 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.286933 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"952372cb-1c5f-4aa9-a751-3526b5ced9a7","Type":"ContainerStarted","Data":"f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb"} Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.295692 4818 generic.go:334] "Generic (PLEG): container finished" podID="ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b" containerID="45c9fba1cd213a75cebfe1bfc9c9756645c2b5a012d0442c0ffee17a3bc40bab" exitCode=0 Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.295783 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b59f8dd-sckpz" event={"ID":"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b","Type":"ContainerDied","Data":"45c9fba1cd213a75cebfe1bfc9c9756645c2b5a012d0442c0ffee17a3bc40bab"} Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.307285 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" event={"ID":"6d8d8f21-1f1f-4431-8079-ffad453ff3f1","Type":"ContainerStarted","Data":"59db56baadb9b4cbf3aebe52d0c59c0888f167bdc5e6054c486e9b067ab50262"} Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.307496 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" event={"ID":"6d8d8f21-1f1f-4431-8079-ffad453ff3f1","Type":"ContainerStarted","Data":"f02d303c1be375bf48a90fc677440a11af6dd95f6f94a54304f6ea23a411c8ed"} Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.309361 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=7.036111395 podStartE2EDuration="8.309343632s" podCreationTimestamp="2025-12-03 06:47:03 +0000 UTC" firstStartedPulling="2025-12-03 06:47:06.424921028 +0000 UTC m=+1184.116529780" lastFinishedPulling="2025-12-03 06:47:07.698153265 +0000 UTC m=+1185.389762017" observedRunningTime="2025-12-03 06:47:11.305104058 +0000 UTC m=+1188.996712810" watchObservedRunningTime="2025-12-03 06:47:11.309343632 +0000 UTC m=+1189.000952384" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.324855 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-554df66f87-jscgb" event={"ID":"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13","Type":"ContainerStarted","Data":"07def945d45670c004fd59488d79a53613a967b40a5d9e93252a33b53944d979"} Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.325058 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-554df66f87-jscgb" event={"ID":"d8e58804-c7d8-48ce-9db3-6c3ab56f1c13","Type":"ContainerStarted","Data":"ee72b1231cd6f10eb06bf73a4040e78171e321799e56eb0a6812ce5c11b977b1"} Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.325398 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-574dc6486d-nvkzv" podUID="bd891b60-48cc-49f1-bb61-9e58c712816e" containerName="horizon-log" containerID="cri-o://1ced693b4186c9669b7fde7dbc2e7323ca419d09850500ca4d98afd3fb031afe" gracePeriod=30 Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.325436 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-574dc6486d-nvkzv" podUID="bd891b60-48cc-49f1-bb61-9e58c712816e" containerName="horizon" containerID="cri-o://b75008d167e0556d0fd21ae4fa3d86d4b00028e3a7b752c00c1288d30868b6b8" gracePeriod=30 Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.328411 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.345134 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.358646 4818 scope.go:117] "RemoveContainer" containerID="1daa47ee7b9c9b5b65ffdd8ef03140fe7f08328ddff57e821221885062ad32bb" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.382206 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7b4848f678-kfvpp" podStartSLOduration=11.023130499 podStartE2EDuration="15.382184449s" podCreationTimestamp="2025-12-03 06:46:56 +0000 UTC" firstStartedPulling="2025-12-03 06:47:06.046542022 +0000 UTC m=+1183.738150774" lastFinishedPulling="2025-12-03 06:47:10.405595972 +0000 UTC m=+1188.097204724" observedRunningTime="2025-12-03 06:47:11.337038841 +0000 UTC m=+1189.028647593" watchObservedRunningTime="2025-12-03 06:47:11.382184449 +0000 UTC m=+1189.073793201" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.403940 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 06:47:11 crc kubenswrapper[4818]: E1203 06:47:11.404327 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f18580e-5450-4edb-b266-734372950182" containerName="horizon-log" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.404343 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f18580e-5450-4edb-b266-734372950182" containerName="horizon-log" Dec 03 06:47:11 crc kubenswrapper[4818]: E1203 06:47:11.404382 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d0cba79-7a13-40ee-ab38-ae093b204a9c" containerName="horizon" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.404388 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d0cba79-7a13-40ee-ab38-ae093b204a9c" containerName="horizon" Dec 03 06:47:11 crc kubenswrapper[4818]: E1203 06:47:11.404396 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="add5819a-c589-4b78-9132-1a8166206a9e" containerName="init" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.404402 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="add5819a-c589-4b78-9132-1a8166206a9e" containerName="init" Dec 03 06:47:11 crc kubenswrapper[4818]: E1203 06:47:11.404411 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d0cba79-7a13-40ee-ab38-ae093b204a9c" containerName="horizon-log" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.404417 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d0cba79-7a13-40ee-ab38-ae093b204a9c" containerName="horizon-log" Dec 03 06:47:11 crc kubenswrapper[4818]: E1203 06:47:11.404430 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc1579d3-d123-45ce-9b92-9486a73de5db" containerName="cinder-api" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.404436 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc1579d3-d123-45ce-9b92-9486a73de5db" containerName="cinder-api" Dec 03 06:47:11 crc kubenswrapper[4818]: E1203 06:47:11.404448 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc1579d3-d123-45ce-9b92-9486a73de5db" containerName="cinder-api-log" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.404454 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc1579d3-d123-45ce-9b92-9486a73de5db" containerName="cinder-api-log" Dec 03 06:47:11 crc kubenswrapper[4818]: E1203 06:47:11.404467 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f18580e-5450-4edb-b266-734372950182" containerName="horizon" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.404472 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f18580e-5450-4edb-b266-734372950182" containerName="horizon" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.404632 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="add5819a-c589-4b78-9132-1a8166206a9e" containerName="init" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.404643 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc1579d3-d123-45ce-9b92-9486a73de5db" containerName="cinder-api-log" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.404658 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d0cba79-7a13-40ee-ab38-ae093b204a9c" containerName="horizon" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.404670 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f18580e-5450-4edb-b266-734372950182" containerName="horizon" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.404681 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc1579d3-d123-45ce-9b92-9486a73de5db" containerName="cinder-api" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.404689 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f18580e-5450-4edb-b266-734372950182" containerName="horizon-log" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.404698 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d0cba79-7a13-40ee-ab38-ae093b204a9c" containerName="horizon-log" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.405659 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.407640 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.407979 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.411355 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-554df66f87-jscgb" podStartSLOduration=11.094347727 podStartE2EDuration="15.411334305s" podCreationTimestamp="2025-12-03 06:46:56 +0000 UTC" firstStartedPulling="2025-12-03 06:47:06.050595191 +0000 UTC m=+1183.742203953" lastFinishedPulling="2025-12-03 06:47:10.367581779 +0000 UTC m=+1188.059190531" observedRunningTime="2025-12-03 06:47:11.387707995 +0000 UTC m=+1189.079316747" watchObservedRunningTime="2025-12-03 06:47:11.411334305 +0000 UTC m=+1189.102943057" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.414438 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.422882 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.457018 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-config-data\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.457099 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/462028f4-61c0-4862-8707-3981513c3b76-logs\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.457186 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.457232 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.457271 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/462028f4-61c0-4862-8707-3981513c3b76-etc-machine-id\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.457360 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds6vb\" (UniqueName: \"kubernetes.io/projected/462028f4-61c0-4862-8707-3981513c3b76-kube-api-access-ds6vb\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.457453 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-scripts\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.457483 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-config-data-custom\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.457591 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-public-tls-certs\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.558407 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/462028f4-61c0-4862-8707-3981513c3b76-logs\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.558733 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.558759 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.559030 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/462028f4-61c0-4862-8707-3981513c3b76-etc-machine-id\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.559524 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds6vb\" (UniqueName: \"kubernetes.io/projected/462028f4-61c0-4862-8707-3981513c3b76-kube-api-access-ds6vb\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.559575 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-scripts\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.559596 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-config-data-custom\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.559641 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-public-tls-certs\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.559669 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-config-data\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.559345 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/462028f4-61c0-4862-8707-3981513c3b76-logs\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.559299 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/462028f4-61c0-4862-8707-3981513c3b76-etc-machine-id\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.565292 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.568504 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-public-tls-certs\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.569206 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-config-data-custom\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.570195 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-scripts\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.570722 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.574770 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/462028f4-61c0-4862-8707-3981513c3b76-config-data\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.589332 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds6vb\" (UniqueName: \"kubernetes.io/projected/462028f4-61c0-4862-8707-3981513c3b76-kube-api-access-ds6vb\") pod \"cinder-api-0\" (UID: \"462028f4-61c0-4862-8707-3981513c3b76\") " pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.757170 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 06:47:11 crc kubenswrapper[4818]: I1203 06:47:11.902230 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.069764 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5d2sg\" (UniqueName: \"kubernetes.io/projected/575fe180-2a23-48e9-b5cf-b90e83528e5e-kube-api-access-5d2sg\") pod \"575fe180-2a23-48e9-b5cf-b90e83528e5e\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.069876 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/575fe180-2a23-48e9-b5cf-b90e83528e5e-run-httpd\") pod \"575fe180-2a23-48e9-b5cf-b90e83528e5e\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.069997 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-combined-ca-bundle\") pod \"575fe180-2a23-48e9-b5cf-b90e83528e5e\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.070054 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/575fe180-2a23-48e9-b5cf-b90e83528e5e-log-httpd\") pod \"575fe180-2a23-48e9-b5cf-b90e83528e5e\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.070078 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-sg-core-conf-yaml\") pod \"575fe180-2a23-48e9-b5cf-b90e83528e5e\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.070096 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-config-data\") pod \"575fe180-2a23-48e9-b5cf-b90e83528e5e\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.070153 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-scripts\") pod \"575fe180-2a23-48e9-b5cf-b90e83528e5e\" (UID: \"575fe180-2a23-48e9-b5cf-b90e83528e5e\") " Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.071496 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/575fe180-2a23-48e9-b5cf-b90e83528e5e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "575fe180-2a23-48e9-b5cf-b90e83528e5e" (UID: "575fe180-2a23-48e9-b5cf-b90e83528e5e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.076291 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/575fe180-2a23-48e9-b5cf-b90e83528e5e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "575fe180-2a23-48e9-b5cf-b90e83528e5e" (UID: "575fe180-2a23-48e9-b5cf-b90e83528e5e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.076624 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/575fe180-2a23-48e9-b5cf-b90e83528e5e-kube-api-access-5d2sg" (OuterVolumeSpecName: "kube-api-access-5d2sg") pod "575fe180-2a23-48e9-b5cf-b90e83528e5e" (UID: "575fe180-2a23-48e9-b5cf-b90e83528e5e"). InnerVolumeSpecName "kube-api-access-5d2sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.076993 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-scripts" (OuterVolumeSpecName: "scripts") pod "575fe180-2a23-48e9-b5cf-b90e83528e5e" (UID: "575fe180-2a23-48e9-b5cf-b90e83528e5e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.104415 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "575fe180-2a23-48e9-b5cf-b90e83528e5e" (UID: "575fe180-2a23-48e9-b5cf-b90e83528e5e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.145948 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "575fe180-2a23-48e9-b5cf-b90e83528e5e" (UID: "575fe180-2a23-48e9-b5cf-b90e83528e5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.173310 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5d2sg\" (UniqueName: \"kubernetes.io/projected/575fe180-2a23-48e9-b5cf-b90e83528e5e-kube-api-access-5d2sg\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.173339 4818 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/575fe180-2a23-48e9-b5cf-b90e83528e5e-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.173348 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.173356 4818 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/575fe180-2a23-48e9-b5cf-b90e83528e5e-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.173364 4818 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.173371 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.204740 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.214188 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-config-data" (OuterVolumeSpecName: "config-data") pod "575fe180-2a23-48e9-b5cf-b90e83528e5e" (UID: "575fe180-2a23-48e9-b5cf-b90e83528e5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.275321 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/575fe180-2a23-48e9-b5cf-b90e83528e5e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.354068 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.360473 4818 generic.go:334] "Generic (PLEG): container finished" podID="575fe180-2a23-48e9-b5cf-b90e83528e5e" containerID="8f699e105a92636cbeab2633821e2e90792849bdc5323a39d6e9a3ee12f8fa9c" exitCode=0 Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.360535 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"575fe180-2a23-48e9-b5cf-b90e83528e5e","Type":"ContainerDied","Data":"8f699e105a92636cbeab2633821e2e90792849bdc5323a39d6e9a3ee12f8fa9c"} Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.360561 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"575fe180-2a23-48e9-b5cf-b90e83528e5e","Type":"ContainerDied","Data":"bc146155042cb9aec31b837131f44bfd6a7e5c6593212a004dcf42473892a271"} Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.360579 4818 scope.go:117] "RemoveContainer" containerID="476e7a3873cbe0683086d8ce5ed7237071f9030a7eedc3126f748c7e3af83718" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.360730 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.375065 4818 generic.go:334] "Generic (PLEG): container finished" podID="ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b" containerID="43fce69ae082d44a241742d1be5d81cb21889f23487c23b323a4b77d499b7453" exitCode=0 Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.375180 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69b59f8dd-sckpz" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.375216 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b59f8dd-sckpz" event={"ID":"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b","Type":"ContainerDied","Data":"43fce69ae082d44a241742d1be5d81cb21889f23487c23b323a4b77d499b7453"} Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.375243 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69b59f8dd-sckpz" event={"ID":"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b","Type":"ContainerDied","Data":"88518310e0da07a729cea5e9ee0a25e8db48bad877849ae2d1daf5d24465f428"} Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.376530 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-httpd-config\") pod \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.376568 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vg6nq\" (UniqueName: \"kubernetes.io/projected/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-kube-api-access-vg6nq\") pod \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.376692 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-combined-ca-bundle\") pod \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.376735 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-config\") pod \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.376798 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-ovndb-tls-certs\") pod \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\" (UID: \"ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b\") " Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.386011 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b" (UID: "ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.388622 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-kube-api-access-vg6nq" (OuterVolumeSpecName: "kube-api-access-vg6nq") pod "ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b" (UID: "ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b"). InnerVolumeSpecName "kube-api-access-vg6nq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.475889 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b" (UID: "ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.480496 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.480642 4818 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.480698 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vg6nq\" (UniqueName: \"kubernetes.io/projected/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-kube-api-access-vg6nq\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.491916 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-config" (OuterVolumeSpecName: "config") pod "ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b" (UID: "ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.492492 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b" (UID: "ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.581934 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.581964 4818 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.600024 4818 scope.go:117] "RemoveContainer" containerID="693c121be8011128371d293e5fc85539a1e5af22221f56702ec76f8eb5db0714" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.675008 4818 scope.go:117] "RemoveContainer" containerID="8f699e105a92636cbeab2633821e2e90792849bdc5323a39d6e9a3ee12f8fa9c" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.679080 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.694686 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.705550 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:12 crc kubenswrapper[4818]: E1203 06:47:12.705922 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="575fe180-2a23-48e9-b5cf-b90e83528e5e" containerName="proxy-httpd" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.705934 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="575fe180-2a23-48e9-b5cf-b90e83528e5e" containerName="proxy-httpd" Dec 03 06:47:12 crc kubenswrapper[4818]: E1203 06:47:12.705959 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b" containerName="neutron-httpd" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.705965 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b" containerName="neutron-httpd" Dec 03 06:47:12 crc kubenswrapper[4818]: E1203 06:47:12.705973 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b" containerName="neutron-api" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.705979 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b" containerName="neutron-api" Dec 03 06:47:12 crc kubenswrapper[4818]: E1203 06:47:12.705992 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="575fe180-2a23-48e9-b5cf-b90e83528e5e" containerName="ceilometer-notification-agent" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.705998 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="575fe180-2a23-48e9-b5cf-b90e83528e5e" containerName="ceilometer-notification-agent" Dec 03 06:47:12 crc kubenswrapper[4818]: E1203 06:47:12.706016 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="575fe180-2a23-48e9-b5cf-b90e83528e5e" containerName="sg-core" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.706022 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="575fe180-2a23-48e9-b5cf-b90e83528e5e" containerName="sg-core" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.706174 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b" containerName="neutron-httpd" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.706188 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b" containerName="neutron-api" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.706197 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="575fe180-2a23-48e9-b5cf-b90e83528e5e" containerName="sg-core" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.706216 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="575fe180-2a23-48e9-b5cf-b90e83528e5e" containerName="proxy-httpd" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.706231 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="575fe180-2a23-48e9-b5cf-b90e83528e5e" containerName="ceilometer-notification-agent" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.707901 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.711004 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.719579 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.740928 4818 scope.go:117] "RemoveContainer" containerID="476e7a3873cbe0683086d8ce5ed7237071f9030a7eedc3126f748c7e3af83718" Dec 03 06:47:12 crc kubenswrapper[4818]: E1203 06:47:12.747220 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"476e7a3873cbe0683086d8ce5ed7237071f9030a7eedc3126f748c7e3af83718\": container with ID starting with 476e7a3873cbe0683086d8ce5ed7237071f9030a7eedc3126f748c7e3af83718 not found: ID does not exist" containerID="476e7a3873cbe0683086d8ce5ed7237071f9030a7eedc3126f748c7e3af83718" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.747382 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"476e7a3873cbe0683086d8ce5ed7237071f9030a7eedc3126f748c7e3af83718"} err="failed to get container status \"476e7a3873cbe0683086d8ce5ed7237071f9030a7eedc3126f748c7e3af83718\": rpc error: code = NotFound desc = could not find container \"476e7a3873cbe0683086d8ce5ed7237071f9030a7eedc3126f748c7e3af83718\": container with ID starting with 476e7a3873cbe0683086d8ce5ed7237071f9030a7eedc3126f748c7e3af83718 not found: ID does not exist" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.747485 4818 scope.go:117] "RemoveContainer" containerID="693c121be8011128371d293e5fc85539a1e5af22221f56702ec76f8eb5db0714" Dec 03 06:47:12 crc kubenswrapper[4818]: E1203 06:47:12.748274 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"693c121be8011128371d293e5fc85539a1e5af22221f56702ec76f8eb5db0714\": container with ID starting with 693c121be8011128371d293e5fc85539a1e5af22221f56702ec76f8eb5db0714 not found: ID does not exist" containerID="693c121be8011128371d293e5fc85539a1e5af22221f56702ec76f8eb5db0714" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.748324 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"693c121be8011128371d293e5fc85539a1e5af22221f56702ec76f8eb5db0714"} err="failed to get container status \"693c121be8011128371d293e5fc85539a1e5af22221f56702ec76f8eb5db0714\": rpc error: code = NotFound desc = could not find container \"693c121be8011128371d293e5fc85539a1e5af22221f56702ec76f8eb5db0714\": container with ID starting with 693c121be8011128371d293e5fc85539a1e5af22221f56702ec76f8eb5db0714 not found: ID does not exist" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.748356 4818 scope.go:117] "RemoveContainer" containerID="8f699e105a92636cbeab2633821e2e90792849bdc5323a39d6e9a3ee12f8fa9c" Dec 03 06:47:12 crc kubenswrapper[4818]: E1203 06:47:12.748760 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f699e105a92636cbeab2633821e2e90792849bdc5323a39d6e9a3ee12f8fa9c\": container with ID starting with 8f699e105a92636cbeab2633821e2e90792849bdc5323a39d6e9a3ee12f8fa9c not found: ID does not exist" containerID="8f699e105a92636cbeab2633821e2e90792849bdc5323a39d6e9a3ee12f8fa9c" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.748792 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f699e105a92636cbeab2633821e2e90792849bdc5323a39d6e9a3ee12f8fa9c"} err="failed to get container status \"8f699e105a92636cbeab2633821e2e90792849bdc5323a39d6e9a3ee12f8fa9c\": rpc error: code = NotFound desc = could not find container \"8f699e105a92636cbeab2633821e2e90792849bdc5323a39d6e9a3ee12f8fa9c\": container with ID starting with 8f699e105a92636cbeab2633821e2e90792849bdc5323a39d6e9a3ee12f8fa9c not found: ID does not exist" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.748848 4818 scope.go:117] "RemoveContainer" containerID="45c9fba1cd213a75cebfe1bfc9c9756645c2b5a012d0442c0ffee17a3bc40bab" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.765073 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="575fe180-2a23-48e9-b5cf-b90e83528e5e" path="/var/lib/kubelet/pods/575fe180-2a23-48e9-b5cf-b90e83528e5e/volumes" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.765761 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc1579d3-d123-45ce-9b92-9486a73de5db" path="/var/lib/kubelet/pods/fc1579d3-d123-45ce-9b92-9486a73de5db/volumes" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.766452 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.766475 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-69b59f8dd-sckpz"] Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.780547 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-69b59f8dd-sckpz"] Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.789638 4818 scope.go:117] "RemoveContainer" containerID="43fce69ae082d44a241742d1be5d81cb21889f23487c23b323a4b77d499b7453" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.834766 4818 scope.go:117] "RemoveContainer" containerID="45c9fba1cd213a75cebfe1bfc9c9756645c2b5a012d0442c0ffee17a3bc40bab" Dec 03 06:47:12 crc kubenswrapper[4818]: E1203 06:47:12.835294 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45c9fba1cd213a75cebfe1bfc9c9756645c2b5a012d0442c0ffee17a3bc40bab\": container with ID starting with 45c9fba1cd213a75cebfe1bfc9c9756645c2b5a012d0442c0ffee17a3bc40bab not found: ID does not exist" containerID="45c9fba1cd213a75cebfe1bfc9c9756645c2b5a012d0442c0ffee17a3bc40bab" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.835337 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45c9fba1cd213a75cebfe1bfc9c9756645c2b5a012d0442c0ffee17a3bc40bab"} err="failed to get container status \"45c9fba1cd213a75cebfe1bfc9c9756645c2b5a012d0442c0ffee17a3bc40bab\": rpc error: code = NotFound desc = could not find container \"45c9fba1cd213a75cebfe1bfc9c9756645c2b5a012d0442c0ffee17a3bc40bab\": container with ID starting with 45c9fba1cd213a75cebfe1bfc9c9756645c2b5a012d0442c0ffee17a3bc40bab not found: ID does not exist" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.835363 4818 scope.go:117] "RemoveContainer" containerID="43fce69ae082d44a241742d1be5d81cb21889f23487c23b323a4b77d499b7453" Dec 03 06:47:12 crc kubenswrapper[4818]: E1203 06:47:12.836176 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43fce69ae082d44a241742d1be5d81cb21889f23487c23b323a4b77d499b7453\": container with ID starting with 43fce69ae082d44a241742d1be5d81cb21889f23487c23b323a4b77d499b7453 not found: ID does not exist" containerID="43fce69ae082d44a241742d1be5d81cb21889f23487c23b323a4b77d499b7453" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.836267 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43fce69ae082d44a241742d1be5d81cb21889f23487c23b323a4b77d499b7453"} err="failed to get container status \"43fce69ae082d44a241742d1be5d81cb21889f23487c23b323a4b77d499b7453\": rpc error: code = NotFound desc = could not find container \"43fce69ae082d44a241742d1be5d81cb21889f23487c23b323a4b77d499b7453\": container with ID starting with 43fce69ae082d44a241742d1be5d81cb21889f23487c23b323a4b77d499b7453 not found: ID does not exist" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.888008 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.888271 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.888424 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-run-httpd\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.888580 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wrgv\" (UniqueName: \"kubernetes.io/projected/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-kube-api-access-2wrgv\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.888788 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-config-data\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.889004 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-scripts\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.889205 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-log-httpd\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.972289 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-748bc8fb48-scr66" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.991349 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wrgv\" (UniqueName: \"kubernetes.io/projected/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-kube-api-access-2wrgv\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.991478 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-config-data\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.991540 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-scripts\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.991579 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-log-httpd\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.991613 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.991636 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.991676 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-run-httpd\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.992190 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-run-httpd\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.993269 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-log-httpd\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:12 crc kubenswrapper[4818]: I1203 06:47:12.997789 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:13 crc kubenswrapper[4818]: I1203 06:47:13.000593 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-config-data\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:13 crc kubenswrapper[4818]: I1203 06:47:13.013675 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:13 crc kubenswrapper[4818]: I1203 06:47:13.014392 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-scripts\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:13 crc kubenswrapper[4818]: I1203 06:47:13.021216 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wrgv\" (UniqueName: \"kubernetes.io/projected/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-kube-api-access-2wrgv\") pod \"ceilometer-0\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " pod="openstack/ceilometer-0" Dec 03 06:47:13 crc kubenswrapper[4818]: I1203 06:47:13.062118 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:47:13 crc kubenswrapper[4818]: I1203 06:47:13.303531 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:47:13 crc kubenswrapper[4818]: I1203 06:47:13.303588 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:47:13 crc kubenswrapper[4818]: I1203 06:47:13.303637 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:47:13 crc kubenswrapper[4818]: I1203 06:47:13.304419 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"94f995293ac99f0a139bded43f6e45522bd84517d494216318f538517e1db600"} pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 06:47:13 crc kubenswrapper[4818]: I1203 06:47:13.304476 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" containerID="cri-o://94f995293ac99f0a139bded43f6e45522bd84517d494216318f538517e1db600" gracePeriod=600 Dec 03 06:47:13 crc kubenswrapper[4818]: I1203 06:47:13.420951 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"462028f4-61c0-4862-8707-3981513c3b76","Type":"ContainerStarted","Data":"6ab4ed38a74a19d685155cba4a77e701cb96c74f18837804c8b0c4e26962489f"} Dec 03 06:47:13 crc kubenswrapper[4818]: I1203 06:47:13.421483 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"462028f4-61c0-4862-8707-3981513c3b76","Type":"ContainerStarted","Data":"207a9488272428512dfaf78b64a08093468255065b5337b37e3aebf0b7c16e84"} Dec 03 06:47:13 crc kubenswrapper[4818]: I1203 06:47:13.799778 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:13 crc kubenswrapper[4818]: I1203 06:47:13.881317 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.004959 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.076624 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-pb59n"] Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.076899 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-pb59n" podUID="8cafb812-b94e-45b2-9f13-9a2be0599382" containerName="dnsmasq-dns" containerID="cri-o://e7b9f5c9eaf99c43a223dbb69a7170a6b6cdb9b5ea6ff0763179bc3f0fbd12ef" gracePeriod=10 Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.274651 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.466372 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15","Type":"ContainerStarted","Data":"3e530a34ce103718704792a86cbb356f9b9d823cb54fc98f0194d6479e858a52"} Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.483166 4818 generic.go:334] "Generic (PLEG): container finished" podID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerID="94f995293ac99f0a139bded43f6e45522bd84517d494216318f538517e1db600" exitCode=0 Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.483275 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerDied","Data":"94f995293ac99f0a139bded43f6e45522bd84517d494216318f538517e1db600"} Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.483352 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"5bf3e0e65704b43280754ba18a3250980330aab9549a78ae7d9381e3d0130a33"} Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.483374 4818 scope.go:117] "RemoveContainer" containerID="9e30bff309bf8bc13ed470ed28234141d81432e279f1c0a4251c4793b659775c" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.488880 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"462028f4-61c0-4862-8707-3981513c3b76","Type":"ContainerStarted","Data":"856b0eda988db5870e597c113e1445d522afb0d82aeec82f284eed6ab5bfa592"} Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.489319 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.503802 4818 generic.go:334] "Generic (PLEG): container finished" podID="8cafb812-b94e-45b2-9f13-9a2be0599382" containerID="e7b9f5c9eaf99c43a223dbb69a7170a6b6cdb9b5ea6ff0763179bc3f0fbd12ef" exitCode=0 Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.505884 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-pb59n" event={"ID":"8cafb812-b94e-45b2-9f13-9a2be0599382","Type":"ContainerDied","Data":"e7b9f5c9eaf99c43a223dbb69a7170a6b6cdb9b5ea6ff0763179bc3f0fbd12ef"} Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.529201 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.529176043 podStartE2EDuration="3.529176043s" podCreationTimestamp="2025-12-03 06:47:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:47:14.522842658 +0000 UTC m=+1192.214451410" watchObservedRunningTime="2025-12-03 06:47:14.529176043 +0000 UTC m=+1192.220784795" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.611209 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.648460 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.650440 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-config\") pod \"8cafb812-b94e-45b2-9f13-9a2be0599382\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.650481 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-ovsdbserver-sb\") pod \"8cafb812-b94e-45b2-9f13-9a2be0599382\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.650508 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-dns-swift-storage-0\") pod \"8cafb812-b94e-45b2-9f13-9a2be0599382\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.650582 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-dns-svc\") pod \"8cafb812-b94e-45b2-9f13-9a2be0599382\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.650742 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-ovsdbserver-nb\") pod \"8cafb812-b94e-45b2-9f13-9a2be0599382\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.650788 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mk7x9\" (UniqueName: \"kubernetes.io/projected/8cafb812-b94e-45b2-9f13-9a2be0599382-kube-api-access-mk7x9\") pod \"8cafb812-b94e-45b2-9f13-9a2be0599382\" (UID: \"8cafb812-b94e-45b2-9f13-9a2be0599382\") " Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.654365 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cafb812-b94e-45b2-9f13-9a2be0599382-kube-api-access-mk7x9" (OuterVolumeSpecName: "kube-api-access-mk7x9") pod "8cafb812-b94e-45b2-9f13-9a2be0599382" (UID: "8cafb812-b94e-45b2-9f13-9a2be0599382"). InnerVolumeSpecName "kube-api-access-mk7x9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.729833 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8cafb812-b94e-45b2-9f13-9a2be0599382" (UID: "8cafb812-b94e-45b2-9f13-9a2be0599382"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.737966 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8cafb812-b94e-45b2-9f13-9a2be0599382" (UID: "8cafb812-b94e-45b2-9f13-9a2be0599382"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.739100 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-config" (OuterVolumeSpecName: "config") pod "8cafb812-b94e-45b2-9f13-9a2be0599382" (UID: "8cafb812-b94e-45b2-9f13-9a2be0599382"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.748498 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b" path="/var/lib/kubelet/pods/ff5cfbd5-5815-4ab8-8d5e-c7afbea33a9b/volumes" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.749149 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8cafb812-b94e-45b2-9f13-9a2be0599382" (UID: "8cafb812-b94e-45b2-9f13-9a2be0599382"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.750364 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8cafb812-b94e-45b2-9f13-9a2be0599382" (UID: "8cafb812-b94e-45b2-9f13-9a2be0599382"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.752771 4818 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.752800 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.752829 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.752842 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mk7x9\" (UniqueName: \"kubernetes.io/projected/8cafb812-b94e-45b2-9f13-9a2be0599382-kube-api-access-mk7x9\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.752855 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.752865 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8cafb812-b94e-45b2-9f13-9a2be0599382-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:14 crc kubenswrapper[4818]: I1203 06:47:14.782676 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6ccc484b58-whz4r" podUID="947dba44-9121-4393-bad5-19439b447c99" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 06:47:15 crc kubenswrapper[4818]: I1203 06:47:15.112450 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:47:15 crc kubenswrapper[4818]: I1203 06:47:15.532154 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-pb59n" event={"ID":"8cafb812-b94e-45b2-9f13-9a2be0599382","Type":"ContainerDied","Data":"5c93308997b1e2bfeb4f6e76ba5f51d6c08c50439e51abbc3bc8c0ecb8a3e200"} Dec 03 06:47:15 crc kubenswrapper[4818]: I1203 06:47:15.532395 4818 scope.go:117] "RemoveContainer" containerID="e7b9f5c9eaf99c43a223dbb69a7170a6b6cdb9b5ea6ff0763179bc3f0fbd12ef" Dec 03 06:47:15 crc kubenswrapper[4818]: I1203 06:47:15.532453 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-pb59n" Dec 03 06:47:15 crc kubenswrapper[4818]: I1203 06:47:15.543983 4818 generic.go:334] "Generic (PLEG): container finished" podID="bd891b60-48cc-49f1-bb61-9e58c712816e" containerID="b75008d167e0556d0fd21ae4fa3d86d4b00028e3a7b752c00c1288d30868b6b8" exitCode=0 Dec 03 06:47:15 crc kubenswrapper[4818]: I1203 06:47:15.544043 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574dc6486d-nvkzv" event={"ID":"bd891b60-48cc-49f1-bb61-9e58c712816e","Type":"ContainerDied","Data":"b75008d167e0556d0fd21ae4fa3d86d4b00028e3a7b752c00c1288d30868b6b8"} Dec 03 06:47:15 crc kubenswrapper[4818]: I1203 06:47:15.554149 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15","Type":"ContainerStarted","Data":"f7e1351a07df87cdcea8bfd1d3a3eb6074ec39ef2685b01fef8bc96adcea4c9f"} Dec 03 06:47:15 crc kubenswrapper[4818]: I1203 06:47:15.554213 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15","Type":"ContainerStarted","Data":"dbe3a64e1f88ab8502e4fd185be00f6da3509d3d4a5bafbe766c26b4b83a2d80"} Dec 03 06:47:15 crc kubenswrapper[4818]: I1203 06:47:15.556199 4818 scope.go:117] "RemoveContainer" containerID="854ceb71d2a1feebf90c5ab71c605729b13e2996f08d6a534983921540014907" Dec 03 06:47:15 crc kubenswrapper[4818]: I1203 06:47:15.569288 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="952372cb-1c5f-4aa9-a751-3526b5ced9a7" containerName="cinder-scheduler" containerID="cri-o://080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3" gracePeriod=30 Dec 03 06:47:15 crc kubenswrapper[4818]: I1203 06:47:15.569880 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="952372cb-1c5f-4aa9-a751-3526b5ced9a7" containerName="probe" containerID="cri-o://f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb" gracePeriod=30 Dec 03 06:47:15 crc kubenswrapper[4818]: I1203 06:47:15.596877 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-pb59n"] Dec 03 06:47:15 crc kubenswrapper[4818]: I1203 06:47:15.620983 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-pb59n"] Dec 03 06:47:15 crc kubenswrapper[4818]: I1203 06:47:15.684930 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-574dc6486d-nvkzv" podUID="bd891b60-48cc-49f1-bb61-9e58c712816e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 03 06:47:16 crc kubenswrapper[4818]: I1203 06:47:16.580636 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15","Type":"ContainerStarted","Data":"66cc1150e8c9b97aec42577c93c3056f251a26faa31ce78caef780f4618bc06f"} Dec 03 06:47:16 crc kubenswrapper[4818]: I1203 06:47:16.747094 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cafb812-b94e-45b2-9f13-9a2be0599382" path="/var/lib/kubelet/pods/8cafb812-b94e-45b2-9f13-9a2be0599382/volumes" Dec 03 06:47:16 crc kubenswrapper[4818]: I1203 06:47:16.939530 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:47:16 crc kubenswrapper[4818]: I1203 06:47:16.941838 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.170986 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7454b897fd-qzp7b" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.245708 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6ccc484b58-whz4r"] Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.408446 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.501744 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-scripts\") pod \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.502000 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf9kj\" (UniqueName: \"kubernetes.io/projected/952372cb-1c5f-4aa9-a751-3526b5ced9a7-kube-api-access-vf9kj\") pod \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.502143 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-config-data-custom\") pod \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.502199 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-config-data\") pod \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.502230 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-combined-ca-bundle\") pod \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.502250 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/952372cb-1c5f-4aa9-a751-3526b5ced9a7-etc-machine-id\") pod \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\" (UID: \"952372cb-1c5f-4aa9-a751-3526b5ced9a7\") " Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.502779 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/952372cb-1c5f-4aa9-a751-3526b5ced9a7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "952372cb-1c5f-4aa9-a751-3526b5ced9a7" (UID: "952372cb-1c5f-4aa9-a751-3526b5ced9a7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.510147 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/952372cb-1c5f-4aa9-a751-3526b5ced9a7-kube-api-access-vf9kj" (OuterVolumeSpecName: "kube-api-access-vf9kj") pod "952372cb-1c5f-4aa9-a751-3526b5ced9a7" (UID: "952372cb-1c5f-4aa9-a751-3526b5ced9a7"). InnerVolumeSpecName "kube-api-access-vf9kj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.522526 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "952372cb-1c5f-4aa9-a751-3526b5ced9a7" (UID: "952372cb-1c5f-4aa9-a751-3526b5ced9a7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.522889 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-scripts" (OuterVolumeSpecName: "scripts") pod "952372cb-1c5f-4aa9-a751-3526b5ced9a7" (UID: "952372cb-1c5f-4aa9-a751-3526b5ced9a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.588954 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "952372cb-1c5f-4aa9-a751-3526b5ced9a7" (UID: "952372cb-1c5f-4aa9-a751-3526b5ced9a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.607159 4818 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.607185 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.607194 4818 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/952372cb-1c5f-4aa9-a751-3526b5ced9a7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.607203 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.607212 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf9kj\" (UniqueName: \"kubernetes.io/projected/952372cb-1c5f-4aa9-a751-3526b5ced9a7-kube-api-access-vf9kj\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.613242 4818 generic.go:334] "Generic (PLEG): container finished" podID="952372cb-1c5f-4aa9-a751-3526b5ced9a7" containerID="f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb" exitCode=0 Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.613272 4818 generic.go:334] "Generic (PLEG): container finished" podID="952372cb-1c5f-4aa9-a751-3526b5ced9a7" containerID="080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3" exitCode=0 Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.613370 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"952372cb-1c5f-4aa9-a751-3526b5ced9a7","Type":"ContainerDied","Data":"f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb"} Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.613437 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"952372cb-1c5f-4aa9-a751-3526b5ced9a7","Type":"ContainerDied","Data":"080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3"} Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.613448 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"952372cb-1c5f-4aa9-a751-3526b5ced9a7","Type":"ContainerDied","Data":"679ef967b9a2abd7f1d98925c44b174d94fbc61e8de7dc02988241b165c24c3c"} Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.613465 4818 scope.go:117] "RemoveContainer" containerID="f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.613655 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.613464 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6ccc484b58-whz4r" podUID="947dba44-9121-4393-bad5-19439b447c99" containerName="barbican-api-log" containerID="cri-o://45399f78c19dec4c370fdf201666cc6eca7f195c8cf3e9c3ec077e9de0263c3d" gracePeriod=30 Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.614544 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6ccc484b58-whz4r" podUID="947dba44-9121-4393-bad5-19439b447c99" containerName="barbican-api" containerID="cri-o://77d24734d359fc520b481b5a57b8bcea5c85b8033ddf54f54f4d688f22546aa9" gracePeriod=30 Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.622157 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6ccc484b58-whz4r" podUID="947dba44-9121-4393-bad5-19439b447c99" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.622286 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6ccc484b58-whz4r" podUID="947dba44-9121-4393-bad5-19439b447c99" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.622457 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6ccc484b58-whz4r" podUID="947dba44-9121-4393-bad5-19439b447c99" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.670969 4818 scope.go:117] "RemoveContainer" containerID="080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.774185 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-config-data" (OuterVolumeSpecName: "config-data") pod "952372cb-1c5f-4aa9-a751-3526b5ced9a7" (UID: "952372cb-1c5f-4aa9-a751-3526b5ced9a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.816043 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/952372cb-1c5f-4aa9-a751-3526b5ced9a7-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.858624 4818 scope.go:117] "RemoveContainer" containerID="f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb" Dec 03 06:47:17 crc kubenswrapper[4818]: E1203 06:47:17.860313 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb\": container with ID starting with f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb not found: ID does not exist" containerID="f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.860475 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb"} err="failed to get container status \"f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb\": rpc error: code = NotFound desc = could not find container \"f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb\": container with ID starting with f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb not found: ID does not exist" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.860594 4818 scope.go:117] "RemoveContainer" containerID="080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3" Dec 03 06:47:17 crc kubenswrapper[4818]: E1203 06:47:17.861166 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3\": container with ID starting with 080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3 not found: ID does not exist" containerID="080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.861205 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3"} err="failed to get container status \"080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3\": rpc error: code = NotFound desc = could not find container \"080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3\": container with ID starting with 080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3 not found: ID does not exist" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.861235 4818 scope.go:117] "RemoveContainer" containerID="f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.861478 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb"} err="failed to get container status \"f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb\": rpc error: code = NotFound desc = could not find container \"f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb\": container with ID starting with f6a701bc01f4d83b58f045e8d8c74db5c57bbb968108f01833f3e4a8a91bd2cb not found: ID does not exist" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.861503 4818 scope.go:117] "RemoveContainer" containerID="080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.861731 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3"} err="failed to get container status \"080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3\": rpc error: code = NotFound desc = could not find container \"080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3\": container with ID starting with 080c50d2084029d286894a888b96178adeaa8c4c4324e2685a391bf437883fe3 not found: ID does not exist" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.956859 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.971431 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.980436 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 06:47:17 crc kubenswrapper[4818]: E1203 06:47:17.981347 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="952372cb-1c5f-4aa9-a751-3526b5ced9a7" containerName="probe" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.981439 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="952372cb-1c5f-4aa9-a751-3526b5ced9a7" containerName="probe" Dec 03 06:47:17 crc kubenswrapper[4818]: E1203 06:47:17.981542 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="952372cb-1c5f-4aa9-a751-3526b5ced9a7" containerName="cinder-scheduler" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.981627 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="952372cb-1c5f-4aa9-a751-3526b5ced9a7" containerName="cinder-scheduler" Dec 03 06:47:17 crc kubenswrapper[4818]: E1203 06:47:17.981715 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cafb812-b94e-45b2-9f13-9a2be0599382" containerName="dnsmasq-dns" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.981874 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cafb812-b94e-45b2-9f13-9a2be0599382" containerName="dnsmasq-dns" Dec 03 06:47:17 crc kubenswrapper[4818]: E1203 06:47:17.981967 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cafb812-b94e-45b2-9f13-9a2be0599382" containerName="init" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.982040 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cafb812-b94e-45b2-9f13-9a2be0599382" containerName="init" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.982326 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cafb812-b94e-45b2-9f13-9a2be0599382" containerName="dnsmasq-dns" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.982427 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="952372cb-1c5f-4aa9-a751-3526b5ced9a7" containerName="probe" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.982520 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="952372cb-1c5f-4aa9-a751-3526b5ced9a7" containerName="cinder-scheduler" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.983778 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.985612 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 06:47:17 crc kubenswrapper[4818]: I1203 06:47:17.990826 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.020184 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ad1caca-cc5b-4625-8776-2fdc829594a4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.020224 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad1caca-cc5b-4625-8776-2fdc829594a4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.020249 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlvcf\" (UniqueName: \"kubernetes.io/projected/5ad1caca-cc5b-4625-8776-2fdc829594a4-kube-api-access-jlvcf\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.020270 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ad1caca-cc5b-4625-8776-2fdc829594a4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.020307 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad1caca-cc5b-4625-8776-2fdc829594a4-config-data\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.020346 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad1caca-cc5b-4625-8776-2fdc829594a4-scripts\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.121356 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad1caca-cc5b-4625-8776-2fdc829594a4-config-data\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.121420 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad1caca-cc5b-4625-8776-2fdc829594a4-scripts\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.121495 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ad1caca-cc5b-4625-8776-2fdc829594a4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.121519 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad1caca-cc5b-4625-8776-2fdc829594a4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.121541 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlvcf\" (UniqueName: \"kubernetes.io/projected/5ad1caca-cc5b-4625-8776-2fdc829594a4-kube-api-access-jlvcf\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.121562 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ad1caca-cc5b-4625-8776-2fdc829594a4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.121632 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ad1caca-cc5b-4625-8776-2fdc829594a4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.129396 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad1caca-cc5b-4625-8776-2fdc829594a4-scripts\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.129700 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ad1caca-cc5b-4625-8776-2fdc829594a4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.130022 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad1caca-cc5b-4625-8776-2fdc829594a4-config-data\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.130646 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad1caca-cc5b-4625-8776-2fdc829594a4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.141120 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlvcf\" (UniqueName: \"kubernetes.io/projected/5ad1caca-cc5b-4625-8776-2fdc829594a4-kube-api-access-jlvcf\") pod \"cinder-scheduler-0\" (UID: \"5ad1caca-cc5b-4625-8776-2fdc829594a4\") " pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.297535 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.628951 4818 generic.go:334] "Generic (PLEG): container finished" podID="947dba44-9121-4393-bad5-19439b447c99" containerID="45399f78c19dec4c370fdf201666cc6eca7f195c8cf3e9c3ec077e9de0263c3d" exitCode=143 Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.629020 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6ccc484b58-whz4r" event={"ID":"947dba44-9121-4393-bad5-19439b447c99","Type":"ContainerDied","Data":"45399f78c19dec4c370fdf201666cc6eca7f195c8cf3e9c3ec077e9de0263c3d"} Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.631686 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15","Type":"ContainerStarted","Data":"352049dac6cf87e85bd53d30232e6676d4849941570daf089d803f18dd05eed4"} Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.633208 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.663152 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.929668842 podStartE2EDuration="6.663133799s" podCreationTimestamp="2025-12-03 06:47:12 +0000 UTC" firstStartedPulling="2025-12-03 06:47:13.819293242 +0000 UTC m=+1191.510901994" lastFinishedPulling="2025-12-03 06:47:17.552758199 +0000 UTC m=+1195.244366951" observedRunningTime="2025-12-03 06:47:18.657964653 +0000 UTC m=+1196.349573415" watchObservedRunningTime="2025-12-03 06:47:18.663133799 +0000 UTC m=+1196.354742561" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.749215 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="952372cb-1c5f-4aa9-a751-3526b5ced9a7" path="/var/lib/kubelet/pods/952372cb-1c5f-4aa9-a751-3526b5ced9a7/volumes" Dec 03 06:47:18 crc kubenswrapper[4818]: I1203 06:47:18.935227 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 06:47:19 crc kubenswrapper[4818]: I1203 06:47:19.643703 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5ad1caca-cc5b-4625-8776-2fdc829594a4","Type":"ContainerStarted","Data":"f6c1dd0256c7196de8ab0ec7dd051f9fe9494cbcdc43dfc944279c522ba5e3a7"} Dec 03 06:47:19 crc kubenswrapper[4818]: I1203 06:47:19.644058 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5ad1caca-cc5b-4625-8776-2fdc829594a4","Type":"ContainerStarted","Data":"8664d653b28d84349e31c6918949996a81a21a5be36ba2cdba73a52f605eafc7"} Dec 03 06:47:20 crc kubenswrapper[4818]: I1203 06:47:20.655554 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5ad1caca-cc5b-4625-8776-2fdc829594a4","Type":"ContainerStarted","Data":"3126534da107c9e841c287bf5bf6e01ce1323177b3fc53ee07f2ace7300567f6"} Dec 03 06:47:20 crc kubenswrapper[4818]: I1203 06:47:20.703911 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.703892224 podStartE2EDuration="3.703892224s" podCreationTimestamp="2025-12-03 06:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:47:20.69843326 +0000 UTC m=+1198.390042042" watchObservedRunningTime="2025-12-03 06:47:20.703892224 +0000 UTC m=+1198.395500976" Dec 03 06:47:22 crc kubenswrapper[4818]: I1203 06:47:22.238724 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-75b5b56458-pbq6h" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.073350 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6ccc484b58-whz4r" podUID="947dba44-9121-4393-bad5-19439b447c99" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:39518->10.217.0.158:9311: read: connection reset by peer" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.073391 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6ccc484b58-whz4r" podUID="947dba44-9121-4393-bad5-19439b447c99" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:39534->10.217.0.158:9311: read: connection reset by peer" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.298770 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.539730 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.584083 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-config-data-custom\") pod \"947dba44-9121-4393-bad5-19439b447c99\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.584179 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-combined-ca-bundle\") pod \"947dba44-9121-4393-bad5-19439b447c99\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.584245 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5fbp\" (UniqueName: \"kubernetes.io/projected/947dba44-9121-4393-bad5-19439b447c99-kube-api-access-k5fbp\") pod \"947dba44-9121-4393-bad5-19439b447c99\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.584285 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-config-data\") pod \"947dba44-9121-4393-bad5-19439b447c99\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.584348 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/947dba44-9121-4393-bad5-19439b447c99-logs\") pod \"947dba44-9121-4393-bad5-19439b447c99\" (UID: \"947dba44-9121-4393-bad5-19439b447c99\") " Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.585267 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/947dba44-9121-4393-bad5-19439b447c99-logs" (OuterVolumeSpecName: "logs") pod "947dba44-9121-4393-bad5-19439b447c99" (UID: "947dba44-9121-4393-bad5-19439b447c99"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.591087 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/947dba44-9121-4393-bad5-19439b447c99-kube-api-access-k5fbp" (OuterVolumeSpecName: "kube-api-access-k5fbp") pod "947dba44-9121-4393-bad5-19439b447c99" (UID: "947dba44-9121-4393-bad5-19439b447c99"). InnerVolumeSpecName "kube-api-access-k5fbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.593284 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "947dba44-9121-4393-bad5-19439b447c99" (UID: "947dba44-9121-4393-bad5-19439b447c99"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.615323 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "947dba44-9121-4393-bad5-19439b447c99" (UID: "947dba44-9121-4393-bad5-19439b447c99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.656777 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-config-data" (OuterVolumeSpecName: "config-data") pod "947dba44-9121-4393-bad5-19439b447c99" (UID: "947dba44-9121-4393-bad5-19439b447c99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.686791 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.686839 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/947dba44-9121-4393-bad5-19439b447c99-logs\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.686849 4818 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.686859 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947dba44-9121-4393-bad5-19439b447c99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.686866 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5fbp\" (UniqueName: \"kubernetes.io/projected/947dba44-9121-4393-bad5-19439b447c99-kube-api-access-k5fbp\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.701389 4818 generic.go:334] "Generic (PLEG): container finished" podID="947dba44-9121-4393-bad5-19439b447c99" containerID="77d24734d359fc520b481b5a57b8bcea5c85b8033ddf54f54f4d688f22546aa9" exitCode=0 Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.701437 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6ccc484b58-whz4r" event={"ID":"947dba44-9121-4393-bad5-19439b447c99","Type":"ContainerDied","Data":"77d24734d359fc520b481b5a57b8bcea5c85b8033ddf54f54f4d688f22546aa9"} Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.701465 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6ccc484b58-whz4r" event={"ID":"947dba44-9121-4393-bad5-19439b447c99","Type":"ContainerDied","Data":"cedf1752781a1598e7d334193cd4a7af0413eee158f699001ae583daa196b62a"} Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.701481 4818 scope.go:117] "RemoveContainer" containerID="77d24734d359fc520b481b5a57b8bcea5c85b8033ddf54f54f4d688f22546aa9" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.701665 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6ccc484b58-whz4r" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.726541 4818 scope.go:117] "RemoveContainer" containerID="45399f78c19dec4c370fdf201666cc6eca7f195c8cf3e9c3ec077e9de0263c3d" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.749770 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6ccc484b58-whz4r"] Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.755681 4818 scope.go:117] "RemoveContainer" containerID="77d24734d359fc520b481b5a57b8bcea5c85b8033ddf54f54f4d688f22546aa9" Dec 03 06:47:23 crc kubenswrapper[4818]: E1203 06:47:23.756588 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77d24734d359fc520b481b5a57b8bcea5c85b8033ddf54f54f4d688f22546aa9\": container with ID starting with 77d24734d359fc520b481b5a57b8bcea5c85b8033ddf54f54f4d688f22546aa9 not found: ID does not exist" containerID="77d24734d359fc520b481b5a57b8bcea5c85b8033ddf54f54f4d688f22546aa9" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.756703 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77d24734d359fc520b481b5a57b8bcea5c85b8033ddf54f54f4d688f22546aa9"} err="failed to get container status \"77d24734d359fc520b481b5a57b8bcea5c85b8033ddf54f54f4d688f22546aa9\": rpc error: code = NotFound desc = could not find container \"77d24734d359fc520b481b5a57b8bcea5c85b8033ddf54f54f4d688f22546aa9\": container with ID starting with 77d24734d359fc520b481b5a57b8bcea5c85b8033ddf54f54f4d688f22546aa9 not found: ID does not exist" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.756782 4818 scope.go:117] "RemoveContainer" containerID="45399f78c19dec4c370fdf201666cc6eca7f195c8cf3e9c3ec077e9de0263c3d" Dec 03 06:47:23 crc kubenswrapper[4818]: E1203 06:47:23.757261 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45399f78c19dec4c370fdf201666cc6eca7f195c8cf3e9c3ec077e9de0263c3d\": container with ID starting with 45399f78c19dec4c370fdf201666cc6eca7f195c8cf3e9c3ec077e9de0263c3d not found: ID does not exist" containerID="45399f78c19dec4c370fdf201666cc6eca7f195c8cf3e9c3ec077e9de0263c3d" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.757298 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45399f78c19dec4c370fdf201666cc6eca7f195c8cf3e9c3ec077e9de0263c3d"} err="failed to get container status \"45399f78c19dec4c370fdf201666cc6eca7f195c8cf3e9c3ec077e9de0263c3d\": rpc error: code = NotFound desc = could not find container \"45399f78c19dec4c370fdf201666cc6eca7f195c8cf3e9c3ec077e9de0263c3d\": container with ID starting with 45399f78c19dec4c370fdf201666cc6eca7f195c8cf3e9c3ec077e9de0263c3d not found: ID does not exist" Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.761486 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6ccc484b58-whz4r"] Dec 03 06:47:23 crc kubenswrapper[4818]: I1203 06:47:23.777603 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 06:47:24 crc kubenswrapper[4818]: I1203 06:47:24.751084 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="947dba44-9121-4393-bad5-19439b447c99" path="/var/lib/kubelet/pods/947dba44-9121-4393-bad5-19439b447c99/volumes" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.370569 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 06:47:25 crc kubenswrapper[4818]: E1203 06:47:25.370921 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="947dba44-9121-4393-bad5-19439b447c99" containerName="barbican-api" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.370936 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="947dba44-9121-4393-bad5-19439b447c99" containerName="barbican-api" Dec 03 06:47:25 crc kubenswrapper[4818]: E1203 06:47:25.370970 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="947dba44-9121-4393-bad5-19439b447c99" containerName="barbican-api-log" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.370979 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="947dba44-9121-4393-bad5-19439b447c99" containerName="barbican-api-log" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.371179 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="947dba44-9121-4393-bad5-19439b447c99" containerName="barbican-api-log" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.371205 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="947dba44-9121-4393-bad5-19439b447c99" containerName="barbican-api" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.371755 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.378186 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-46jf9" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.378392 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.378502 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.391661 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.518685 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aecf49dc-5a72-41aa-bc7c-c42bf226ea71-openstack-config-secret\") pod \"openstackclient\" (UID: \"aecf49dc-5a72-41aa-bc7c-c42bf226ea71\") " pod="openstack/openstackclient" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.518734 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aecf49dc-5a72-41aa-bc7c-c42bf226ea71-combined-ca-bundle\") pod \"openstackclient\" (UID: \"aecf49dc-5a72-41aa-bc7c-c42bf226ea71\") " pod="openstack/openstackclient" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.518858 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjbg7\" (UniqueName: \"kubernetes.io/projected/aecf49dc-5a72-41aa-bc7c-c42bf226ea71-kube-api-access-pjbg7\") pod \"openstackclient\" (UID: \"aecf49dc-5a72-41aa-bc7c-c42bf226ea71\") " pod="openstack/openstackclient" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.518900 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aecf49dc-5a72-41aa-bc7c-c42bf226ea71-openstack-config\") pod \"openstackclient\" (UID: \"aecf49dc-5a72-41aa-bc7c-c42bf226ea71\") " pod="openstack/openstackclient" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.620444 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjbg7\" (UniqueName: \"kubernetes.io/projected/aecf49dc-5a72-41aa-bc7c-c42bf226ea71-kube-api-access-pjbg7\") pod \"openstackclient\" (UID: \"aecf49dc-5a72-41aa-bc7c-c42bf226ea71\") " pod="openstack/openstackclient" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.620496 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aecf49dc-5a72-41aa-bc7c-c42bf226ea71-openstack-config\") pod \"openstackclient\" (UID: \"aecf49dc-5a72-41aa-bc7c-c42bf226ea71\") " pod="openstack/openstackclient" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.620555 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aecf49dc-5a72-41aa-bc7c-c42bf226ea71-openstack-config-secret\") pod \"openstackclient\" (UID: \"aecf49dc-5a72-41aa-bc7c-c42bf226ea71\") " pod="openstack/openstackclient" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.620586 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aecf49dc-5a72-41aa-bc7c-c42bf226ea71-combined-ca-bundle\") pod \"openstackclient\" (UID: \"aecf49dc-5a72-41aa-bc7c-c42bf226ea71\") " pod="openstack/openstackclient" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.622984 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aecf49dc-5a72-41aa-bc7c-c42bf226ea71-openstack-config\") pod \"openstackclient\" (UID: \"aecf49dc-5a72-41aa-bc7c-c42bf226ea71\") " pod="openstack/openstackclient" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.626376 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aecf49dc-5a72-41aa-bc7c-c42bf226ea71-openstack-config-secret\") pod \"openstackclient\" (UID: \"aecf49dc-5a72-41aa-bc7c-c42bf226ea71\") " pod="openstack/openstackclient" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.626676 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aecf49dc-5a72-41aa-bc7c-c42bf226ea71-combined-ca-bundle\") pod \"openstackclient\" (UID: \"aecf49dc-5a72-41aa-bc7c-c42bf226ea71\") " pod="openstack/openstackclient" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.636773 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjbg7\" (UniqueName: \"kubernetes.io/projected/aecf49dc-5a72-41aa-bc7c-c42bf226ea71-kube-api-access-pjbg7\") pod \"openstackclient\" (UID: \"aecf49dc-5a72-41aa-bc7c-c42bf226ea71\") " pod="openstack/openstackclient" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.685044 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-574dc6486d-nvkzv" podUID="bd891b60-48cc-49f1-bb61-9e58c712816e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 03 06:47:25 crc kubenswrapper[4818]: I1203 06:47:25.703911 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 06:47:26 crc kubenswrapper[4818]: I1203 06:47:26.169939 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 06:47:26 crc kubenswrapper[4818]: I1203 06:47:26.730148 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"aecf49dc-5a72-41aa-bc7c-c42bf226ea71","Type":"ContainerStarted","Data":"e36f292ebd8490a73a3b08828a46a2721188607b4a1c4415c4a57a65404cd8e1"} Dec 03 06:47:28 crc kubenswrapper[4818]: I1203 06:47:28.562090 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 06:47:30 crc kubenswrapper[4818]: I1203 06:47:30.708985 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:30 crc kubenswrapper[4818]: I1203 06:47:30.709729 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerName="ceilometer-central-agent" containerID="cri-o://dbe3a64e1f88ab8502e4fd185be00f6da3509d3d4a5bafbe766c26b4b83a2d80" gracePeriod=30 Dec 03 06:47:30 crc kubenswrapper[4818]: I1203 06:47:30.710605 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerName="proxy-httpd" containerID="cri-o://352049dac6cf87e85bd53d30232e6676d4849941570daf089d803f18dd05eed4" gracePeriod=30 Dec 03 06:47:30 crc kubenswrapper[4818]: I1203 06:47:30.710669 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerName="ceilometer-notification-agent" containerID="cri-o://f7e1351a07df87cdcea8bfd1d3a3eb6074ec39ef2685b01fef8bc96adcea4c9f" gracePeriod=30 Dec 03 06:47:30 crc kubenswrapper[4818]: I1203 06:47:30.710796 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerName="sg-core" containerID="cri-o://66cc1150e8c9b97aec42577c93c3056f251a26faa31ce78caef780f4618bc06f" gracePeriod=30 Dec 03 06:47:30 crc kubenswrapper[4818]: I1203 06:47:30.719594 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.164:3000/\": EOF" Dec 03 06:47:30 crc kubenswrapper[4818]: I1203 06:47:30.869052 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7954d7ccff-th9lw"] Dec 03 06:47:30 crc kubenswrapper[4818]: I1203 06:47:30.871015 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:30 crc kubenswrapper[4818]: I1203 06:47:30.879708 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 03 06:47:30 crc kubenswrapper[4818]: I1203 06:47:30.880023 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 03 06:47:30 crc kubenswrapper[4818]: I1203 06:47:30.880227 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 06:47:30 crc kubenswrapper[4818]: I1203 06:47:30.885290 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7954d7ccff-th9lw"] Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.032671 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjqvf\" (UniqueName: \"kubernetes.io/projected/0ba50d38-959d-49f4-a98a-6280a78661e4-kube-api-access-pjqvf\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.032748 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0ba50d38-959d-49f4-a98a-6280a78661e4-etc-swift\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.032779 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba50d38-959d-49f4-a98a-6280a78661e4-internal-tls-certs\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.032800 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ba50d38-959d-49f4-a98a-6280a78661e4-log-httpd\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.032877 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ba50d38-959d-49f4-a98a-6280a78661e4-run-httpd\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.032951 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba50d38-959d-49f4-a98a-6280a78661e4-public-tls-certs\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.033008 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ba50d38-959d-49f4-a98a-6280a78661e4-config-data\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.033029 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba50d38-959d-49f4-a98a-6280a78661e4-combined-ca-bundle\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.134673 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ba50d38-959d-49f4-a98a-6280a78661e4-run-httpd\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.134790 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba50d38-959d-49f4-a98a-6280a78661e4-public-tls-certs\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.134965 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ba50d38-959d-49f4-a98a-6280a78661e4-config-data\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.134995 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba50d38-959d-49f4-a98a-6280a78661e4-combined-ca-bundle\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.135037 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjqvf\" (UniqueName: \"kubernetes.io/projected/0ba50d38-959d-49f4-a98a-6280a78661e4-kube-api-access-pjqvf\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.135069 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0ba50d38-959d-49f4-a98a-6280a78661e4-etc-swift\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.135102 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba50d38-959d-49f4-a98a-6280a78661e4-internal-tls-certs\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.135130 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ba50d38-959d-49f4-a98a-6280a78661e4-log-httpd\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.135302 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ba50d38-959d-49f4-a98a-6280a78661e4-run-httpd\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.135566 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ba50d38-959d-49f4-a98a-6280a78661e4-log-httpd\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.143929 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba50d38-959d-49f4-a98a-6280a78661e4-combined-ca-bundle\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.143986 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba50d38-959d-49f4-a98a-6280a78661e4-internal-tls-certs\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.144260 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ba50d38-959d-49f4-a98a-6280a78661e4-config-data\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.144326 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba50d38-959d-49f4-a98a-6280a78661e4-public-tls-certs\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.145549 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0ba50d38-959d-49f4-a98a-6280a78661e4-etc-swift\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.161134 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjqvf\" (UniqueName: \"kubernetes.io/projected/0ba50d38-959d-49f4-a98a-6280a78661e4-kube-api-access-pjqvf\") pod \"swift-proxy-7954d7ccff-th9lw\" (UID: \"0ba50d38-959d-49f4-a98a-6280a78661e4\") " pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.207236 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.786906 4818 generic.go:334] "Generic (PLEG): container finished" podID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerID="352049dac6cf87e85bd53d30232e6676d4849941570daf089d803f18dd05eed4" exitCode=0 Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.786943 4818 generic.go:334] "Generic (PLEG): container finished" podID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerID="66cc1150e8c9b97aec42577c93c3056f251a26faa31ce78caef780f4618bc06f" exitCode=2 Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.786954 4818 generic.go:334] "Generic (PLEG): container finished" podID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerID="dbe3a64e1f88ab8502e4fd185be00f6da3509d3d4a5bafbe766c26b4b83a2d80" exitCode=0 Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.786976 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15","Type":"ContainerDied","Data":"352049dac6cf87e85bd53d30232e6676d4849941570daf089d803f18dd05eed4"} Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.787020 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15","Type":"ContainerDied","Data":"66cc1150e8c9b97aec42577c93c3056f251a26faa31ce78caef780f4618bc06f"} Dec 03 06:47:31 crc kubenswrapper[4818]: I1203 06:47:31.787030 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15","Type":"ContainerDied","Data":"dbe3a64e1f88ab8502e4fd185be00f6da3509d3d4a5bafbe766c26b4b83a2d80"} Dec 03 06:47:33 crc kubenswrapper[4818]: I1203 06:47:33.808463 4818 generic.go:334] "Generic (PLEG): container finished" podID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerID="f7e1351a07df87cdcea8bfd1d3a3eb6074ec39ef2685b01fef8bc96adcea4c9f" exitCode=0 Dec 03 06:47:33 crc kubenswrapper[4818]: I1203 06:47:33.808874 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15","Type":"ContainerDied","Data":"f7e1351a07df87cdcea8bfd1d3a3eb6074ec39ef2685b01fef8bc96adcea4c9f"} Dec 03 06:47:35 crc kubenswrapper[4818]: I1203 06:47:35.684604 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-574dc6486d-nvkzv" podUID="bd891b60-48cc-49f1-bb61-9e58c712816e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 03 06:47:35 crc kubenswrapper[4818]: I1203 06:47:35.684994 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:47:35 crc kubenswrapper[4818]: I1203 06:47:35.737188 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 06:47:35 crc kubenswrapper[4818]: I1203 06:47:35.738163 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="34a77ce6-1705-45d6-af7f-8100e13a6d3b" containerName="glance-log" containerID="cri-o://a463adb758a277f6ea665085b981251f1e9da331319b2d1f8de125625611894b" gracePeriod=30 Dec 03 06:47:35 crc kubenswrapper[4818]: I1203 06:47:35.738307 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="34a77ce6-1705-45d6-af7f-8100e13a6d3b" containerName="glance-httpd" containerID="cri-o://31516935f93bba37172720180afa57d9aa0dbfa9449a6a0543bde59e2a83b036" gracePeriod=30 Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.840137 4818 generic.go:334] "Generic (PLEG): container finished" podID="34a77ce6-1705-45d6-af7f-8100e13a6d3b" containerID="a463adb758a277f6ea665085b981251f1e9da331319b2d1f8de125625611894b" exitCode=143 Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.841026 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"34a77ce6-1705-45d6-af7f-8100e13a6d3b","Type":"ContainerDied","Data":"a463adb758a277f6ea665085b981251f1e9da331319b2d1f8de125625611894b"} Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.841765 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.848051 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15","Type":"ContainerDied","Data":"3e530a34ce103718704792a86cbb356f9b9d823cb54fc98f0194d6479e858a52"} Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.848102 4818 scope.go:117] "RemoveContainer" containerID="352049dac6cf87e85bd53d30232e6676d4849941570daf089d803f18dd05eed4" Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.886679 4818 scope.go:117] "RemoveContainer" containerID="66cc1150e8c9b97aec42577c93c3056f251a26faa31ce78caef780f4618bc06f" Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.913955 4818 scope.go:117] "RemoveContainer" containerID="f7e1351a07df87cdcea8bfd1d3a3eb6074ec39ef2685b01fef8bc96adcea4c9f" Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.940411 4818 scope.go:117] "RemoveContainer" containerID="dbe3a64e1f88ab8502e4fd185be00f6da3509d3d4a5bafbe766c26b4b83a2d80" Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.963202 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-log-httpd\") pod \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.963304 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-scripts\") pod \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.963402 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-config-data\") pod \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.963529 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wrgv\" (UniqueName: \"kubernetes.io/projected/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-kube-api-access-2wrgv\") pod \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.963571 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-combined-ca-bundle\") pod \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.963644 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-run-httpd\") pod \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.963679 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-sg-core-conf-yaml\") pod \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\" (UID: \"341b0bb6-ed9e-44af-9e2b-182ff4cf8d15\") " Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.964048 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" (UID: "341b0bb6-ed9e-44af-9e2b-182ff4cf8d15"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.964244 4818 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.969034 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" (UID: "341b0bb6-ed9e-44af-9e2b-182ff4cf8d15"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.969663 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-kube-api-access-2wrgv" (OuterVolumeSpecName: "kube-api-access-2wrgv") pod "341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" (UID: "341b0bb6-ed9e-44af-9e2b-182ff4cf8d15"). InnerVolumeSpecName "kube-api-access-2wrgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:36 crc kubenswrapper[4818]: I1203 06:47:36.982438 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-scripts" (OuterVolumeSpecName: "scripts") pod "341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" (UID: "341b0bb6-ed9e-44af-9e2b-182ff4cf8d15"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.007389 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" (UID: "341b0bb6-ed9e-44af-9e2b-182ff4cf8d15"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.063382 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" (UID: "341b0bb6-ed9e-44af-9e2b-182ff4cf8d15"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.064900 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wrgv\" (UniqueName: \"kubernetes.io/projected/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-kube-api-access-2wrgv\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.064923 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.064933 4818 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.064941 4818 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.064949 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.089545 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-config-data" (OuterVolumeSpecName: "config-data") pod "341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" (UID: "341b0bb6-ed9e-44af-9e2b-182ff4cf8d15"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.165969 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.290032 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7954d7ccff-th9lw"] Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.866544 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"aecf49dc-5a72-41aa-bc7c-c42bf226ea71","Type":"ContainerStarted","Data":"0c8e18a526e87178a00918436370d469881587e29f41a9da57f972f9bb1721c4"} Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.868788 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.873669 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7954d7ccff-th9lw" event={"ID":"0ba50d38-959d-49f4-a98a-6280a78661e4","Type":"ContainerStarted","Data":"f33a9ab9d482713ffa2bcb5db177725b0677758135e807169fe2b44c407d807c"} Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.873713 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7954d7ccff-th9lw" event={"ID":"0ba50d38-959d-49f4-a98a-6280a78661e4","Type":"ContainerStarted","Data":"8b228f23f8cd7cb9ea903c027ded4ed85fa5c3d2c2a79e308c5f06904fec132c"} Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.873726 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7954d7ccff-th9lw" event={"ID":"0ba50d38-959d-49f4-a98a-6280a78661e4","Type":"ContainerStarted","Data":"edfd2e77a62268a68e84c9ca984bb13b625ff06628124051b8529db38fae532b"} Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.874518 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.874556 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.899155 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.479651976 podStartE2EDuration="12.899134452s" podCreationTimestamp="2025-12-03 06:47:25 +0000 UTC" firstStartedPulling="2025-12-03 06:47:26.175027057 +0000 UTC m=+1203.866635809" lastFinishedPulling="2025-12-03 06:47:36.594509533 +0000 UTC m=+1214.286118285" observedRunningTime="2025-12-03 06:47:37.888801688 +0000 UTC m=+1215.580410430" watchObservedRunningTime="2025-12-03 06:47:37.899134452 +0000 UTC m=+1215.590743204" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.918703 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.935544 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.945012 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7954d7ccff-th9lw" podStartSLOduration=7.944989857 podStartE2EDuration="7.944989857s" podCreationTimestamp="2025-12-03 06:47:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:47:37.927200191 +0000 UTC m=+1215.618808943" watchObservedRunningTime="2025-12-03 06:47:37.944989857 +0000 UTC m=+1215.636598609" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.955023 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:37 crc kubenswrapper[4818]: E1203 06:47:37.955420 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerName="sg-core" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.955439 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerName="sg-core" Dec 03 06:47:37 crc kubenswrapper[4818]: E1203 06:47:37.955457 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerName="ceilometer-central-agent" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.955465 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerName="ceilometer-central-agent" Dec 03 06:47:37 crc kubenswrapper[4818]: E1203 06:47:37.955479 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerName="ceilometer-notification-agent" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.955486 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerName="ceilometer-notification-agent" Dec 03 06:47:37 crc kubenswrapper[4818]: E1203 06:47:37.955496 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerName="proxy-httpd" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.955502 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerName="proxy-httpd" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.955669 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerName="ceilometer-central-agent" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.955684 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerName="ceilometer-notification-agent" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.955694 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerName="sg-core" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.955706 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" containerName="proxy-httpd" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.957397 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.959617 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.959928 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.969960 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.981993 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-config-data\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.982332 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-scripts\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.982626 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84520fa4-ba04-4475-ab4f-ab20f9ddc870-run-httpd\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.982661 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.982722 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.982794 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84520fa4-ba04-4475-ab4f-ab20f9ddc870-log-httpd\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:37 crc kubenswrapper[4818]: I1203 06:47:37.982920 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4xbk\" (UniqueName: \"kubernetes.io/projected/84520fa4-ba04-4475-ab4f-ab20f9ddc870-kube-api-access-r4xbk\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.084810 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84520fa4-ba04-4475-ab4f-ab20f9ddc870-run-httpd\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.084884 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.084926 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.084999 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84520fa4-ba04-4475-ab4f-ab20f9ddc870-log-httpd\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.085151 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4xbk\" (UniqueName: \"kubernetes.io/projected/84520fa4-ba04-4475-ab4f-ab20f9ddc870-kube-api-access-r4xbk\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.085186 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-config-data\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.085610 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84520fa4-ba04-4475-ab4f-ab20f9ddc870-run-httpd\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.085717 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84520fa4-ba04-4475-ab4f-ab20f9ddc870-log-httpd\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.085780 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-scripts\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.091091 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.091249 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.092682 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-scripts\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.094005 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-config-data\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.104594 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4xbk\" (UniqueName: \"kubernetes.io/projected/84520fa4-ba04-4475-ab4f-ab20f9ddc870-kube-api-access-r4xbk\") pod \"ceilometer-0\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " pod="openstack/ceilometer-0" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.284394 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.554297 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-vn2ml"] Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.561759 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vn2ml" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.580742 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-vn2ml"] Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.644377 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.644622 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" containerName="glance-log" containerID="cri-o://f5d9e63e248172bb123c2119799b48c42c47dc27abf1ccde1c22c8db92974276" gracePeriod=30 Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.644716 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" containerName="glance-httpd" containerID="cri-o://d741ce548f75e3da7cf0a8bd816e096a18f0c6cec4097ed3fe4b1cb7da83e0aa" gracePeriod=30 Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.699065 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlm9v\" (UniqueName: \"kubernetes.io/projected/ae575e67-4d14-4f4c-8d27-b74f9b837134-kube-api-access-wlm9v\") pod \"nova-api-db-create-vn2ml\" (UID: \"ae575e67-4d14-4f4c-8d27-b74f9b837134\") " pod="openstack/nova-api-db-create-vn2ml" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.699116 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae575e67-4d14-4f4c-8d27-b74f9b837134-operator-scripts\") pod \"nova-api-db-create-vn2ml\" (UID: \"ae575e67-4d14-4f4c-8d27-b74f9b837134\") " pod="openstack/nova-api-db-create-vn2ml" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.751260 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="341b0bb6-ed9e-44af-9e2b-182ff4cf8d15" path="/var/lib/kubelet/pods/341b0bb6-ed9e-44af-9e2b-182ff4cf8d15/volumes" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.752787 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-sb9lv"] Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.757032 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-sb9lv" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.800495 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlm9v\" (UniqueName: \"kubernetes.io/projected/ae575e67-4d14-4f4c-8d27-b74f9b837134-kube-api-access-wlm9v\") pod \"nova-api-db-create-vn2ml\" (UID: \"ae575e67-4d14-4f4c-8d27-b74f9b837134\") " pod="openstack/nova-api-db-create-vn2ml" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.800547 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae575e67-4d14-4f4c-8d27-b74f9b837134-operator-scripts\") pod \"nova-api-db-create-vn2ml\" (UID: \"ae575e67-4d14-4f4c-8d27-b74f9b837134\") " pod="openstack/nova-api-db-create-vn2ml" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.801263 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae575e67-4d14-4f4c-8d27-b74f9b837134-operator-scripts\") pod \"nova-api-db-create-vn2ml\" (UID: \"ae575e67-4d14-4f4c-8d27-b74f9b837134\") " pod="openstack/nova-api-db-create-vn2ml" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.820687 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-sb9lv"] Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.827024 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlm9v\" (UniqueName: \"kubernetes.io/projected/ae575e67-4d14-4f4c-8d27-b74f9b837134-kube-api-access-wlm9v\") pod \"nova-api-db-create-vn2ml\" (UID: \"ae575e67-4d14-4f4c-8d27-b74f9b837134\") " pod="openstack/nova-api-db-create-vn2ml" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.834099 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-8ffc-account-create-update-s65jt"] Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.835293 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8ffc-account-create-update-s65jt" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.837086 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.867971 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8ffc-account-create-update-s65jt"] Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.875403 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.881879 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84520fa4-ba04-4475-ab4f-ab20f9ddc870","Type":"ContainerStarted","Data":"445ca1e4e0c204f3c98228839f07fec9341ab2027b736636e9fdcb695560c9e9"} Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.882511 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-gff9c"] Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.883654 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gff9c" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.885533 4818 generic.go:334] "Generic (PLEG): container finished" podID="11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" containerID="f5d9e63e248172bb123c2119799b48c42c47dc27abf1ccde1c22c8db92974276" exitCode=143 Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.885760 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f","Type":"ContainerDied","Data":"f5d9e63e248172bb123c2119799b48c42c47dc27abf1ccde1c22c8db92974276"} Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.893105 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vn2ml" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.902462 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdbn2\" (UniqueName: \"kubernetes.io/projected/5184be52-db4d-4e95-88bd-a6569ddb75ff-kube-api-access-zdbn2\") pod \"nova-cell0-db-create-sb9lv\" (UID: \"5184be52-db4d-4e95-88bd-a6569ddb75ff\") " pod="openstack/nova-cell0-db-create-sb9lv" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.902570 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5184be52-db4d-4e95-88bd-a6569ddb75ff-operator-scripts\") pod \"nova-cell0-db-create-sb9lv\" (UID: \"5184be52-db4d-4e95-88bd-a6569ddb75ff\") " pod="openstack/nova-cell0-db-create-sb9lv" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.908191 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-gff9c"] Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.987669 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-e5ab-account-create-update-n6pl8"] Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.988992 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e5ab-account-create-update-n6pl8" Dec 03 06:47:38 crc kubenswrapper[4818]: I1203 06:47:38.994121 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.012346 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvc2n\" (UniqueName: \"kubernetes.io/projected/f2194aad-43ff-4ab6-b648-e079a0688a5f-kube-api-access-fvc2n\") pod \"nova-cell0-e5ab-account-create-update-n6pl8\" (UID: \"f2194aad-43ff-4ab6-b648-e079a0688a5f\") " pod="openstack/nova-cell0-e5ab-account-create-update-n6pl8" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.012400 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5184be52-db4d-4e95-88bd-a6569ddb75ff-operator-scripts\") pod \"nova-cell0-db-create-sb9lv\" (UID: \"5184be52-db4d-4e95-88bd-a6569ddb75ff\") " pod="openstack/nova-cell0-db-create-sb9lv" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.012456 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f0e2585-56c1-435c-9db7-74b19a1cab80-operator-scripts\") pod \"nova-api-8ffc-account-create-update-s65jt\" (UID: \"7f0e2585-56c1-435c-9db7-74b19a1cab80\") " pod="openstack/nova-api-8ffc-account-create-update-s65jt" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.012551 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdbn2\" (UniqueName: \"kubernetes.io/projected/5184be52-db4d-4e95-88bd-a6569ddb75ff-kube-api-access-zdbn2\") pod \"nova-cell0-db-create-sb9lv\" (UID: \"5184be52-db4d-4e95-88bd-a6569ddb75ff\") " pod="openstack/nova-cell0-db-create-sb9lv" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.012603 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7514750-68e9-462c-b401-f80689f8f4df-operator-scripts\") pod \"nova-cell1-db-create-gff9c\" (UID: \"f7514750-68e9-462c-b401-f80689f8f4df\") " pod="openstack/nova-cell1-db-create-gff9c" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.012629 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtvs7\" (UniqueName: \"kubernetes.io/projected/7f0e2585-56c1-435c-9db7-74b19a1cab80-kube-api-access-jtvs7\") pod \"nova-api-8ffc-account-create-update-s65jt\" (UID: \"7f0e2585-56c1-435c-9db7-74b19a1cab80\") " pod="openstack/nova-api-8ffc-account-create-update-s65jt" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.012653 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjq9l\" (UniqueName: \"kubernetes.io/projected/f7514750-68e9-462c-b401-f80689f8f4df-kube-api-access-pjq9l\") pod \"nova-cell1-db-create-gff9c\" (UID: \"f7514750-68e9-462c-b401-f80689f8f4df\") " pod="openstack/nova-cell1-db-create-gff9c" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.012704 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2194aad-43ff-4ab6-b648-e079a0688a5f-operator-scripts\") pod \"nova-cell0-e5ab-account-create-update-n6pl8\" (UID: \"f2194aad-43ff-4ab6-b648-e079a0688a5f\") " pod="openstack/nova-cell0-e5ab-account-create-update-n6pl8" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.012364 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-e5ab-account-create-update-n6pl8"] Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.014191 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5184be52-db4d-4e95-88bd-a6569ddb75ff-operator-scripts\") pod \"nova-cell0-db-create-sb9lv\" (UID: \"5184be52-db4d-4e95-88bd-a6569ddb75ff\") " pod="openstack/nova-cell0-db-create-sb9lv" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.048686 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdbn2\" (UniqueName: \"kubernetes.io/projected/5184be52-db4d-4e95-88bd-a6569ddb75ff-kube-api-access-zdbn2\") pod \"nova-cell0-db-create-sb9lv\" (UID: \"5184be52-db4d-4e95-88bd-a6569ddb75ff\") " pod="openstack/nova-cell0-db-create-sb9lv" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.113997 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvc2n\" (UniqueName: \"kubernetes.io/projected/f2194aad-43ff-4ab6-b648-e079a0688a5f-kube-api-access-fvc2n\") pod \"nova-cell0-e5ab-account-create-update-n6pl8\" (UID: \"f2194aad-43ff-4ab6-b648-e079a0688a5f\") " pod="openstack/nova-cell0-e5ab-account-create-update-n6pl8" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.114287 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f0e2585-56c1-435c-9db7-74b19a1cab80-operator-scripts\") pod \"nova-api-8ffc-account-create-update-s65jt\" (UID: \"7f0e2585-56c1-435c-9db7-74b19a1cab80\") " pod="openstack/nova-api-8ffc-account-create-update-s65jt" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.114376 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7514750-68e9-462c-b401-f80689f8f4df-operator-scripts\") pod \"nova-cell1-db-create-gff9c\" (UID: \"f7514750-68e9-462c-b401-f80689f8f4df\") " pod="openstack/nova-cell1-db-create-gff9c" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.114397 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtvs7\" (UniqueName: \"kubernetes.io/projected/7f0e2585-56c1-435c-9db7-74b19a1cab80-kube-api-access-jtvs7\") pod \"nova-api-8ffc-account-create-update-s65jt\" (UID: \"7f0e2585-56c1-435c-9db7-74b19a1cab80\") " pod="openstack/nova-api-8ffc-account-create-update-s65jt" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.114418 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjq9l\" (UniqueName: \"kubernetes.io/projected/f7514750-68e9-462c-b401-f80689f8f4df-kube-api-access-pjq9l\") pod \"nova-cell1-db-create-gff9c\" (UID: \"f7514750-68e9-462c-b401-f80689f8f4df\") " pod="openstack/nova-cell1-db-create-gff9c" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.114456 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2194aad-43ff-4ab6-b648-e079a0688a5f-operator-scripts\") pod \"nova-cell0-e5ab-account-create-update-n6pl8\" (UID: \"f2194aad-43ff-4ab6-b648-e079a0688a5f\") " pod="openstack/nova-cell0-e5ab-account-create-update-n6pl8" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.115393 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2194aad-43ff-4ab6-b648-e079a0688a5f-operator-scripts\") pod \"nova-cell0-e5ab-account-create-update-n6pl8\" (UID: \"f2194aad-43ff-4ab6-b648-e079a0688a5f\") " pod="openstack/nova-cell0-e5ab-account-create-update-n6pl8" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.116154 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f0e2585-56c1-435c-9db7-74b19a1cab80-operator-scripts\") pod \"nova-api-8ffc-account-create-update-s65jt\" (UID: \"7f0e2585-56c1-435c-9db7-74b19a1cab80\") " pod="openstack/nova-api-8ffc-account-create-update-s65jt" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.116881 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7514750-68e9-462c-b401-f80689f8f4df-operator-scripts\") pod \"nova-cell1-db-create-gff9c\" (UID: \"f7514750-68e9-462c-b401-f80689f8f4df\") " pod="openstack/nova-cell1-db-create-gff9c" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.162932 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvc2n\" (UniqueName: \"kubernetes.io/projected/f2194aad-43ff-4ab6-b648-e079a0688a5f-kube-api-access-fvc2n\") pod \"nova-cell0-e5ab-account-create-update-n6pl8\" (UID: \"f2194aad-43ff-4ab6-b648-e079a0688a5f\") " pod="openstack/nova-cell0-e5ab-account-create-update-n6pl8" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.168729 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjq9l\" (UniqueName: \"kubernetes.io/projected/f7514750-68e9-462c-b401-f80689f8f4df-kube-api-access-pjq9l\") pod \"nova-cell1-db-create-gff9c\" (UID: \"f7514750-68e9-462c-b401-f80689f8f4df\") " pod="openstack/nova-cell1-db-create-gff9c" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.169450 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gff9c" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.172485 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtvs7\" (UniqueName: \"kubernetes.io/projected/7f0e2585-56c1-435c-9db7-74b19a1cab80-kube-api-access-jtvs7\") pod \"nova-api-8ffc-account-create-update-s65jt\" (UID: \"7f0e2585-56c1-435c-9db7-74b19a1cab80\") " pod="openstack/nova-api-8ffc-account-create-update-s65jt" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.185875 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-5631-account-create-update-k87bh"] Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.187003 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5631-account-create-update-k87bh" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.209919 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.220769 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e5ab-account-create-update-n6pl8" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.236874 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5631-account-create-update-k87bh"] Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.260209 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-sb9lv" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.319321 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5bps\" (UniqueName: \"kubernetes.io/projected/4dcf3310-6518-4c25-a81c-ef30724db1b8-kube-api-access-g5bps\") pod \"nova-cell1-5631-account-create-update-k87bh\" (UID: \"4dcf3310-6518-4c25-a81c-ef30724db1b8\") " pod="openstack/nova-cell1-5631-account-create-update-k87bh" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.319392 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dcf3310-6518-4c25-a81c-ef30724db1b8-operator-scripts\") pod \"nova-cell1-5631-account-create-update-k87bh\" (UID: \"4dcf3310-6518-4c25-a81c-ef30724db1b8\") " pod="openstack/nova-cell1-5631-account-create-update-k87bh" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.421365 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5bps\" (UniqueName: \"kubernetes.io/projected/4dcf3310-6518-4c25-a81c-ef30724db1b8-kube-api-access-g5bps\") pod \"nova-cell1-5631-account-create-update-k87bh\" (UID: \"4dcf3310-6518-4c25-a81c-ef30724db1b8\") " pod="openstack/nova-cell1-5631-account-create-update-k87bh" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.421750 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dcf3310-6518-4c25-a81c-ef30724db1b8-operator-scripts\") pod \"nova-cell1-5631-account-create-update-k87bh\" (UID: \"4dcf3310-6518-4c25-a81c-ef30724db1b8\") " pod="openstack/nova-cell1-5631-account-create-update-k87bh" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.423968 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dcf3310-6518-4c25-a81c-ef30724db1b8-operator-scripts\") pod \"nova-cell1-5631-account-create-update-k87bh\" (UID: \"4dcf3310-6518-4c25-a81c-ef30724db1b8\") " pod="openstack/nova-cell1-5631-account-create-update-k87bh" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.435094 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8ffc-account-create-update-s65jt" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.459749 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5bps\" (UniqueName: \"kubernetes.io/projected/4dcf3310-6518-4c25-a81c-ef30724db1b8-kube-api-access-g5bps\") pod \"nova-cell1-5631-account-create-update-k87bh\" (UID: \"4dcf3310-6518-4c25-a81c-ef30724db1b8\") " pod="openstack/nova-cell1-5631-account-create-update-k87bh" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.544242 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5631-account-create-update-k87bh" Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.779107 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-vn2ml"] Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.909731 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84520fa4-ba04-4475-ab4f-ab20f9ddc870","Type":"ContainerStarted","Data":"938292b21c559eca7770c1f64eae322064ebadab0d0155422f47e22e0dce950f"} Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.913691 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-vn2ml" event={"ID":"ae575e67-4d14-4f4c-8d27-b74f9b837134","Type":"ContainerStarted","Data":"f2060496997429091873a09eac0c94b8e1bdc0b92a9daa4ab4c0a231a7cb38bf"} Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.939387 4818 generic.go:334] "Generic (PLEG): container finished" podID="34a77ce6-1705-45d6-af7f-8100e13a6d3b" containerID="31516935f93bba37172720180afa57d9aa0dbfa9449a6a0543bde59e2a83b036" exitCode=0 Dec 03 06:47:39 crc kubenswrapper[4818]: I1203 06:47:39.941299 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"34a77ce6-1705-45d6-af7f-8100e13a6d3b","Type":"ContainerDied","Data":"31516935f93bba37172720180afa57d9aa0dbfa9449a6a0543bde59e2a83b036"} Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.089000 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.159246 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-e5ab-account-create-update-n6pl8"] Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.167777 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-gff9c"] Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.244862 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-public-tls-certs\") pod \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.244920 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-scripts\") pod \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.245029 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-config-data\") pod \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.245077 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/34a77ce6-1705-45d6-af7f-8100e13a6d3b-httpd-run\") pod \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.245247 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-combined-ca-bundle\") pod \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.245306 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.245453 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58r5k\" (UniqueName: \"kubernetes.io/projected/34a77ce6-1705-45d6-af7f-8100e13a6d3b-kube-api-access-58r5k\") pod \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.245485 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34a77ce6-1705-45d6-af7f-8100e13a6d3b-logs\") pod \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\" (UID: \"34a77ce6-1705-45d6-af7f-8100e13a6d3b\") " Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.246374 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34a77ce6-1705-45d6-af7f-8100e13a6d3b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "34a77ce6-1705-45d6-af7f-8100e13a6d3b" (UID: "34a77ce6-1705-45d6-af7f-8100e13a6d3b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.246777 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34a77ce6-1705-45d6-af7f-8100e13a6d3b-logs" (OuterVolumeSpecName: "logs") pod "34a77ce6-1705-45d6-af7f-8100e13a6d3b" (UID: "34a77ce6-1705-45d6-af7f-8100e13a6d3b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.254494 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34a77ce6-1705-45d6-af7f-8100e13a6d3b-kube-api-access-58r5k" (OuterVolumeSpecName: "kube-api-access-58r5k") pod "34a77ce6-1705-45d6-af7f-8100e13a6d3b" (UID: "34a77ce6-1705-45d6-af7f-8100e13a6d3b"). InnerVolumeSpecName "kube-api-access-58r5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.257982 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "34a77ce6-1705-45d6-af7f-8100e13a6d3b" (UID: "34a77ce6-1705-45d6-af7f-8100e13a6d3b"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.258114 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-scripts" (OuterVolumeSpecName: "scripts") pod "34a77ce6-1705-45d6-af7f-8100e13a6d3b" (UID: "34a77ce6-1705-45d6-af7f-8100e13a6d3b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.310965 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34a77ce6-1705-45d6-af7f-8100e13a6d3b" (UID: "34a77ce6-1705-45d6-af7f-8100e13a6d3b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.339371 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "34a77ce6-1705-45d6-af7f-8100e13a6d3b" (UID: "34a77ce6-1705-45d6-af7f-8100e13a6d3b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.350952 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58r5k\" (UniqueName: \"kubernetes.io/projected/34a77ce6-1705-45d6-af7f-8100e13a6d3b-kube-api-access-58r5k\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.350984 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34a77ce6-1705-45d6-af7f-8100e13a6d3b-logs\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.350992 4818 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.351000 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.351007 4818 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/34a77ce6-1705-45d6-af7f-8100e13a6d3b-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.351015 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.351047 4818 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.395176 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5631-account-create-update-k87bh"] Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.457284 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8ffc-account-create-update-s65jt"] Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.482370 4818 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.533708 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-config-data" (OuterVolumeSpecName: "config-data") pod "34a77ce6-1705-45d6-af7f-8100e13a6d3b" (UID: "34a77ce6-1705-45d6-af7f-8100e13a6d3b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.538379 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-sb9lv"] Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.568976 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a77ce6-1705-45d6-af7f-8100e13a6d3b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.569013 4818 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.671620 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.957291 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-sb9lv" event={"ID":"5184be52-db4d-4e95-88bd-a6569ddb75ff","Type":"ContainerStarted","Data":"9a5b024c89797b5d8e12e17458eacb212be6237dfb792e9eaab96ad014dd7a11"} Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.960041 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"34a77ce6-1705-45d6-af7f-8100e13a6d3b","Type":"ContainerDied","Data":"7ba1a8fab1700ed6f9676c759f75dc7bf7d9ca4bb8dcf07630cb26aa3bd0c287"} Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.960122 4818 scope.go:117] "RemoveContainer" containerID="31516935f93bba37172720180afa57d9aa0dbfa9449a6a0543bde59e2a83b036" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.960338 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.965142 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8ffc-account-create-update-s65jt" event={"ID":"7f0e2585-56c1-435c-9db7-74b19a1cab80","Type":"ContainerStarted","Data":"48df0b580e2fa868b7da60dfc4768d0c1dbe9032758462364322868f06cbcd54"} Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.970714 4818 generic.go:334] "Generic (PLEG): container finished" podID="f2194aad-43ff-4ab6-b648-e079a0688a5f" containerID="7aa75430e5eda89fc452183734b53c46d491e15429f34cbebdaf233a74322f2c" exitCode=0 Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.971156 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e5ab-account-create-update-n6pl8" event={"ID":"f2194aad-43ff-4ab6-b648-e079a0688a5f","Type":"ContainerDied","Data":"7aa75430e5eda89fc452183734b53c46d491e15429f34cbebdaf233a74322f2c"} Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.971196 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e5ab-account-create-update-n6pl8" event={"ID":"f2194aad-43ff-4ab6-b648-e079a0688a5f","Type":"ContainerStarted","Data":"1bba4812ce83b4f47c35763b1c297053c8dba6c72912d9c616e6e43a3d4adcc5"} Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.979715 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5631-account-create-update-k87bh" event={"ID":"4dcf3310-6518-4c25-a81c-ef30724db1b8","Type":"ContainerStarted","Data":"b60d29a02e8947cf9dc5db637aaacce93e351eec8db851fa064f755a2b15e6a6"} Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.989609 4818 generic.go:334] "Generic (PLEG): container finished" podID="f7514750-68e9-462c-b401-f80689f8f4df" containerID="c38ac13a89962c7d38a2a09e6bc8c242c29b7d8e01fba581a8988df80d6820fd" exitCode=0 Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.989676 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-gff9c" event={"ID":"f7514750-68e9-462c-b401-f80689f8f4df","Type":"ContainerDied","Data":"c38ac13a89962c7d38a2a09e6bc8c242c29b7d8e01fba581a8988df80d6820fd"} Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.989748 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-gff9c" event={"ID":"f7514750-68e9-462c-b401-f80689f8f4df","Type":"ContainerStarted","Data":"f6e82c1181103f771a0dd902372a100677ae5bb6984b73ee51ed45247e82eea9"} Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.989884 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.994186 4818 generic.go:334] "Generic (PLEG): container finished" podID="ae575e67-4d14-4f4c-8d27-b74f9b837134" containerID="40a20a4a921d35cbe1c425b32780cb643b7e62e79caade8cedeed1463cf89cf3" exitCode=0 Dec 03 06:47:40 crc kubenswrapper[4818]: I1203 06:47:40.994219 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-vn2ml" event={"ID":"ae575e67-4d14-4f4c-8d27-b74f9b837134","Type":"ContainerDied","Data":"40a20a4a921d35cbe1c425b32780cb643b7e62e79caade8cedeed1463cf89cf3"} Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.005544 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.019719 4818 scope.go:117] "RemoveContainer" containerID="a463adb758a277f6ea665085b981251f1e9da331319b2d1f8de125625611894b" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.027668 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 06:47:41 crc kubenswrapper[4818]: E1203 06:47:41.028132 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34a77ce6-1705-45d6-af7f-8100e13a6d3b" containerName="glance-httpd" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.028157 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="34a77ce6-1705-45d6-af7f-8100e13a6d3b" containerName="glance-httpd" Dec 03 06:47:41 crc kubenswrapper[4818]: E1203 06:47:41.028183 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34a77ce6-1705-45d6-af7f-8100e13a6d3b" containerName="glance-log" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.028190 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="34a77ce6-1705-45d6-af7f-8100e13a6d3b" containerName="glance-log" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.028411 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="34a77ce6-1705-45d6-af7f-8100e13a6d3b" containerName="glance-log" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.028438 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="34a77ce6-1705-45d6-af7f-8100e13a6d3b" containerName="glance-httpd" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.033529 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.040936 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.041116 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.050197 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.180233 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c46a9a7-6f86-40b8-8957-5d0f11d11262-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.180279 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c46a9a7-6f86-40b8-8957-5d0f11d11262-logs\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.180755 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.180879 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c46a9a7-6f86-40b8-8957-5d0f11d11262-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.180924 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbxwl\" (UniqueName: \"kubernetes.io/projected/0c46a9a7-6f86-40b8-8957-5d0f11d11262-kube-api-access-mbxwl\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.181024 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c46a9a7-6f86-40b8-8957-5d0f11d11262-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.181089 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c46a9a7-6f86-40b8-8957-5d0f11d11262-scripts\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.181290 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c46a9a7-6f86-40b8-8957-5d0f11d11262-config-data\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.286682 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c46a9a7-6f86-40b8-8957-5d0f11d11262-config-data\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.286742 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c46a9a7-6f86-40b8-8957-5d0f11d11262-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.286763 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c46a9a7-6f86-40b8-8957-5d0f11d11262-logs\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.286870 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.286894 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c46a9a7-6f86-40b8-8957-5d0f11d11262-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.286918 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbxwl\" (UniqueName: \"kubernetes.io/projected/0c46a9a7-6f86-40b8-8957-5d0f11d11262-kube-api-access-mbxwl\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.286947 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c46a9a7-6f86-40b8-8957-5d0f11d11262-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.286966 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c46a9a7-6f86-40b8-8957-5d0f11d11262-scripts\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.287385 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c46a9a7-6f86-40b8-8957-5d0f11d11262-logs\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.288070 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.288512 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c46a9a7-6f86-40b8-8957-5d0f11d11262-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.295146 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c46a9a7-6f86-40b8-8957-5d0f11d11262-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.297964 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c46a9a7-6f86-40b8-8957-5d0f11d11262-config-data\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.298624 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c46a9a7-6f86-40b8-8957-5d0f11d11262-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.304560 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c46a9a7-6f86-40b8-8957-5d0f11d11262-scripts\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.319472 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbxwl\" (UniqueName: \"kubernetes.io/projected/0c46a9a7-6f86-40b8-8957-5d0f11d11262-kube-api-access-mbxwl\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.345040 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"0c46a9a7-6f86-40b8-8957-5d0f11d11262\") " pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: I1203 06:47:41.365680 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 06:47:41 crc kubenswrapper[4818]: E1203 06:47:41.897172 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11dff550_2d3d_4bbc_bbf8_e1bb6849c94f.slice/crio-conmon-f5d9e63e248172bb123c2119799b48c42c47dc27abf1ccde1c22c8db92974276.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34a77ce6_1705_45d6_af7f_8100e13a6d3b.slice/crio-31516935f93bba37172720180afa57d9aa0dbfa9449a6a0543bde59e2a83b036.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34a77ce6_1705_45d6_af7f_8100e13a6d3b.slice/crio-conmon-31516935f93bba37172720180afa57d9aa0dbfa9449a6a0543bde59e2a83b036.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd891b60_48cc_49f1_bb61_9e58c712816e.slice/crio-conmon-1ced693b4186c9669b7fde7dbc2e7323ca419d09850500ca4d98afd3fb031afe.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34a77ce6_1705_45d6_af7f_8100e13a6d3b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34a77ce6_1705_45d6_af7f_8100e13a6d3b.slice/crio-7ba1a8fab1700ed6f9676c759f75dc7bf7d9ca4bb8dcf07630cb26aa3bd0c287\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11dff550_2d3d_4bbc_bbf8_e1bb6849c94f.slice/crio-f5d9e63e248172bb123c2119799b48c42c47dc27abf1ccde1c22c8db92974276.scope\": RecentStats: unable to find data in memory cache]" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.012391 4818 generic.go:334] "Generic (PLEG): container finished" podID="7f0e2585-56c1-435c-9db7-74b19a1cab80" containerID="c02300f0d9f614a680c6a19b0d6e6dddc1cd095065722b3b46911227341903fa" exitCode=0 Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.012462 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8ffc-account-create-update-s65jt" event={"ID":"7f0e2585-56c1-435c-9db7-74b19a1cab80","Type":"ContainerDied","Data":"c02300f0d9f614a680c6a19b0d6e6dddc1cd095065722b3b46911227341903fa"} Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.015367 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84520fa4-ba04-4475-ab4f-ab20f9ddc870","Type":"ContainerStarted","Data":"dc4b5dcb248545aaf154ff7ee4db1271573719e5f4d9a5b34d9c617e48bb27bb"} Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.017008 4818 generic.go:334] "Generic (PLEG): container finished" podID="bd891b60-48cc-49f1-bb61-9e58c712816e" containerID="1ced693b4186c9669b7fde7dbc2e7323ca419d09850500ca4d98afd3fb031afe" exitCode=137 Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.017075 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574dc6486d-nvkzv" event={"ID":"bd891b60-48cc-49f1-bb61-9e58c712816e","Type":"ContainerDied","Data":"1ced693b4186c9669b7fde7dbc2e7323ca419d09850500ca4d98afd3fb031afe"} Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.017092 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574dc6486d-nvkzv" event={"ID":"bd891b60-48cc-49f1-bb61-9e58c712816e","Type":"ContainerDied","Data":"40aec6581deeb30113d6bd25b29b2eb16765b914a7708455f81e07f5041b88c0"} Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.017101 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40aec6581deeb30113d6bd25b29b2eb16765b914a7708455f81e07f5041b88c0" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.022842 4818 generic.go:334] "Generic (PLEG): container finished" podID="11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" containerID="d741ce548f75e3da7cf0a8bd816e096a18f0c6cec4097ed3fe4b1cb7da83e0aa" exitCode=0 Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.023053 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f","Type":"ContainerDied","Data":"d741ce548f75e3da7cf0a8bd816e096a18f0c6cec4097ed3fe4b1cb7da83e0aa"} Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.024457 4818 generic.go:334] "Generic (PLEG): container finished" podID="4dcf3310-6518-4c25-a81c-ef30724db1b8" containerID="74684d24944f4d5aed5bbb34bd9cd0d54f3d12dc4525efa4bb85050dd7694d97" exitCode=0 Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.024536 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5631-account-create-update-k87bh" event={"ID":"4dcf3310-6518-4c25-a81c-ef30724db1b8","Type":"ContainerDied","Data":"74684d24944f4d5aed5bbb34bd9cd0d54f3d12dc4525efa4bb85050dd7694d97"} Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.027085 4818 generic.go:334] "Generic (PLEG): container finished" podID="5184be52-db4d-4e95-88bd-a6569ddb75ff" containerID="2051b0fea1f9046bee62c7be1d04a2562c80f88640717a85b180d8513376c4e0" exitCode=0 Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.027173 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-sb9lv" event={"ID":"5184be52-db4d-4e95-88bd-a6569ddb75ff","Type":"ContainerDied","Data":"2051b0fea1f9046bee62c7be1d04a2562c80f88640717a85b180d8513376c4e0"} Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.066040 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.231360 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-horizon-tls-certs\") pod \"bd891b60-48cc-49f1-bb61-9e58c712816e\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.231506 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-horizon-secret-key\") pod \"bd891b60-48cc-49f1-bb61-9e58c712816e\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.231540 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68pvc\" (UniqueName: \"kubernetes.io/projected/bd891b60-48cc-49f1-bb61-9e58c712816e-kube-api-access-68pvc\") pod \"bd891b60-48cc-49f1-bb61-9e58c712816e\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.231570 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd891b60-48cc-49f1-bb61-9e58c712816e-scripts\") pod \"bd891b60-48cc-49f1-bb61-9e58c712816e\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.231643 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-combined-ca-bundle\") pod \"bd891b60-48cc-49f1-bb61-9e58c712816e\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.231661 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd891b60-48cc-49f1-bb61-9e58c712816e-logs\") pod \"bd891b60-48cc-49f1-bb61-9e58c712816e\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.231704 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd891b60-48cc-49f1-bb61-9e58c712816e-config-data\") pod \"bd891b60-48cc-49f1-bb61-9e58c712816e\" (UID: \"bd891b60-48cc-49f1-bb61-9e58c712816e\") " Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.236681 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd891b60-48cc-49f1-bb61-9e58c712816e-logs" (OuterVolumeSpecName: "logs") pod "bd891b60-48cc-49f1-bb61-9e58c712816e" (UID: "bd891b60-48cc-49f1-bb61-9e58c712816e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.244460 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd891b60-48cc-49f1-bb61-9e58c712816e-kube-api-access-68pvc" (OuterVolumeSpecName: "kube-api-access-68pvc") pod "bd891b60-48cc-49f1-bb61-9e58c712816e" (UID: "bd891b60-48cc-49f1-bb61-9e58c712816e"). InnerVolumeSpecName "kube-api-access-68pvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.257041 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "bd891b60-48cc-49f1-bb61-9e58c712816e" (UID: "bd891b60-48cc-49f1-bb61-9e58c712816e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.285319 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd891b60-48cc-49f1-bb61-9e58c712816e-scripts" (OuterVolumeSpecName: "scripts") pod "bd891b60-48cc-49f1-bb61-9e58c712816e" (UID: "bd891b60-48cc-49f1-bb61-9e58c712816e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.334141 4818 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.334188 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68pvc\" (UniqueName: \"kubernetes.io/projected/bd891b60-48cc-49f1-bb61-9e58c712816e-kube-api-access-68pvc\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.334232 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd891b60-48cc-49f1-bb61-9e58c712816e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.334243 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd891b60-48cc-49f1-bb61-9e58c712816e-logs\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.352032 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd891b60-48cc-49f1-bb61-9e58c712816e" (UID: "bd891b60-48cc-49f1-bb61-9e58c712816e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.390696 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "bd891b60-48cc-49f1-bb61-9e58c712816e" (UID: "bd891b60-48cc-49f1-bb61-9e58c712816e"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.420723 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd891b60-48cc-49f1-bb61-9e58c712816e-config-data" (OuterVolumeSpecName: "config-data") pod "bd891b60-48cc-49f1-bb61-9e58c712816e" (UID: "bd891b60-48cc-49f1-bb61-9e58c712816e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.455303 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.455350 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd891b60-48cc-49f1-bb61-9e58c712816e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.455362 4818 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd891b60-48cc-49f1-bb61-9e58c712816e-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.655343 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.763762 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-config-data\") pod \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.763910 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-combined-ca-bundle\") pod \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.763964 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g4mg\" (UniqueName: \"kubernetes.io/projected/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-kube-api-access-6g4mg\") pod \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.764011 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.764032 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-scripts\") pod \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.764039 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34a77ce6-1705-45d6-af7f-8100e13a6d3b" path="/var/lib/kubelet/pods/34a77ce6-1705-45d6-af7f-8100e13a6d3b/volumes" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.787792 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-kube-api-access-6g4mg" (OuterVolumeSpecName: "kube-api-access-6g4mg") pod "11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" (UID: "11dff550-2d3d-4bbc-bbf8-e1bb6849c94f"). InnerVolumeSpecName "kube-api-access-6g4mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.764060 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-internal-tls-certs\") pod \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.788465 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-httpd-run\") pod \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.788599 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-logs\") pod \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\" (UID: \"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f\") " Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.789867 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g4mg\" (UniqueName: \"kubernetes.io/projected/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-kube-api-access-6g4mg\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.791205 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" (UID: "11dff550-2d3d-4bbc-bbf8-e1bb6849c94f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.794311 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" (UID: "11dff550-2d3d-4bbc-bbf8-e1bb6849c94f"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.794607 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-logs" (OuterVolumeSpecName: "logs") pod "11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" (UID: "11dff550-2d3d-4bbc-bbf8-e1bb6849c94f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.811987 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-scripts" (OuterVolumeSpecName: "scripts") pod "11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" (UID: "11dff550-2d3d-4bbc-bbf8-e1bb6849c94f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.861965 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" (UID: "11dff550-2d3d-4bbc-bbf8-e1bb6849c94f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.892219 4818 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.892270 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.892281 4818 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.892292 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-logs\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.892305 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.894903 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" (UID: "11dff550-2d3d-4bbc-bbf8-e1bb6849c94f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:42 crc kubenswrapper[4818]: I1203 06:47:42.996034 4818 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.018987 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-config-data" (OuterVolumeSpecName: "config-data") pod "11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" (UID: "11dff550-2d3d-4bbc-bbf8-e1bb6849c94f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.034759 4818 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.047287 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e5ab-account-create-update-n6pl8" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.052603 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gff9c" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.053431 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-gff9c" event={"ID":"f7514750-68e9-462c-b401-f80689f8f4df","Type":"ContainerDied","Data":"f6e82c1181103f771a0dd902372a100677ae5bb6984b73ee51ed45247e82eea9"} Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.053486 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6e82c1181103f771a0dd902372a100677ae5bb6984b73ee51ed45247e82eea9" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.056972 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-vn2ml" event={"ID":"ae575e67-4d14-4f4c-8d27-b74f9b837134","Type":"ContainerDied","Data":"f2060496997429091873a09eac0c94b8e1bdc0b92a9daa4ab4c0a231a7cb38bf"} Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.057125 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2060496997429091873a09eac0c94b8e1bdc0b92a9daa4ab4c0a231a7cb38bf" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.059455 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vn2ml" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.060275 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e5ab-account-create-update-n6pl8" event={"ID":"f2194aad-43ff-4ab6-b648-e079a0688a5f","Type":"ContainerDied","Data":"1bba4812ce83b4f47c35763b1c297053c8dba6c72912d9c616e6e43a3d4adcc5"} Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.060330 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bba4812ce83b4f47c35763b1c297053c8dba6c72912d9c616e6e43a3d4adcc5" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.060386 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e5ab-account-create-update-n6pl8" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.078263 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84520fa4-ba04-4475-ab4f-ab20f9ddc870","Type":"ContainerStarted","Data":"fc8f2925aa00371b84d68ec50ba5127addda07934af03dcc5db1cb124364777f"} Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.083005 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.083465 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-574dc6486d-nvkzv" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.084523 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"11dff550-2d3d-4bbc-bbf8-e1bb6849c94f","Type":"ContainerDied","Data":"6ffe09976590646045227ddb478e5fd8d01c9354f24f473bc74dabce43fdd282"} Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.084591 4818 scope.go:117] "RemoveContainer" containerID="d741ce548f75e3da7cf0a8bd816e096a18f0c6cec4097ed3fe4b1cb7da83e0aa" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.098372 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjq9l\" (UniqueName: \"kubernetes.io/projected/f7514750-68e9-462c-b401-f80689f8f4df-kube-api-access-pjq9l\") pod \"f7514750-68e9-462c-b401-f80689f8f4df\" (UID: \"f7514750-68e9-462c-b401-f80689f8f4df\") " Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.098572 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlm9v\" (UniqueName: \"kubernetes.io/projected/ae575e67-4d14-4f4c-8d27-b74f9b837134-kube-api-access-wlm9v\") pod \"ae575e67-4d14-4f4c-8d27-b74f9b837134\" (UID: \"ae575e67-4d14-4f4c-8d27-b74f9b837134\") " Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.098745 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2194aad-43ff-4ab6-b648-e079a0688a5f-operator-scripts\") pod \"f2194aad-43ff-4ab6-b648-e079a0688a5f\" (UID: \"f2194aad-43ff-4ab6-b648-e079a0688a5f\") " Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.099062 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae575e67-4d14-4f4c-8d27-b74f9b837134-operator-scripts\") pod \"ae575e67-4d14-4f4c-8d27-b74f9b837134\" (UID: \"ae575e67-4d14-4f4c-8d27-b74f9b837134\") " Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.099091 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7514750-68e9-462c-b401-f80689f8f4df-operator-scripts\") pod \"f7514750-68e9-462c-b401-f80689f8f4df\" (UID: \"f7514750-68e9-462c-b401-f80689f8f4df\") " Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.099152 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvc2n\" (UniqueName: \"kubernetes.io/projected/f2194aad-43ff-4ab6-b648-e079a0688a5f-kube-api-access-fvc2n\") pod \"f2194aad-43ff-4ab6-b648-e079a0688a5f\" (UID: \"f2194aad-43ff-4ab6-b648-e079a0688a5f\") " Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.100066 4818 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.100083 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.103621 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2194aad-43ff-4ab6-b648-e079a0688a5f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f2194aad-43ff-4ab6-b648-e079a0688a5f" (UID: "f2194aad-43ff-4ab6-b648-e079a0688a5f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.104058 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae575e67-4d14-4f4c-8d27-b74f9b837134-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ae575e67-4d14-4f4c-8d27-b74f9b837134" (UID: "ae575e67-4d14-4f4c-8d27-b74f9b837134"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.104520 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7514750-68e9-462c-b401-f80689f8f4df-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f7514750-68e9-462c-b401-f80689f8f4df" (UID: "f7514750-68e9-462c-b401-f80689f8f4df"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.115046 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.117355 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7514750-68e9-462c-b401-f80689f8f4df-kube-api-access-pjq9l" (OuterVolumeSpecName: "kube-api-access-pjq9l") pod "f7514750-68e9-462c-b401-f80689f8f4df" (UID: "f7514750-68e9-462c-b401-f80689f8f4df"). InnerVolumeSpecName "kube-api-access-pjq9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.120701 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2194aad-43ff-4ab6-b648-e079a0688a5f-kube-api-access-fvc2n" (OuterVolumeSpecName: "kube-api-access-fvc2n") pod "f2194aad-43ff-4ab6-b648-e079a0688a5f" (UID: "f2194aad-43ff-4ab6-b648-e079a0688a5f"). InnerVolumeSpecName "kube-api-access-fvc2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.128794 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae575e67-4d14-4f4c-8d27-b74f9b837134-kube-api-access-wlm9v" (OuterVolumeSpecName: "kube-api-access-wlm9v") pod "ae575e67-4d14-4f4c-8d27-b74f9b837134" (UID: "ae575e67-4d14-4f4c-8d27-b74f9b837134"). InnerVolumeSpecName "kube-api-access-wlm9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.149886 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-574dc6486d-nvkzv"] Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.159833 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-574dc6486d-nvkzv"] Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.166272 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.177027 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.185199 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 06:47:43 crc kubenswrapper[4818]: E1203 06:47:43.185561 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd891b60-48cc-49f1-bb61-9e58c712816e" containerName="horizon-log" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.185578 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd891b60-48cc-49f1-bb61-9e58c712816e" containerName="horizon-log" Dec 03 06:47:43 crc kubenswrapper[4818]: E1203 06:47:43.185600 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" containerName="glance-httpd" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.185607 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" containerName="glance-httpd" Dec 03 06:47:43 crc kubenswrapper[4818]: E1203 06:47:43.185617 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2194aad-43ff-4ab6-b648-e079a0688a5f" containerName="mariadb-account-create-update" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.185623 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2194aad-43ff-4ab6-b648-e079a0688a5f" containerName="mariadb-account-create-update" Dec 03 06:47:43 crc kubenswrapper[4818]: E1203 06:47:43.185639 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7514750-68e9-462c-b401-f80689f8f4df" containerName="mariadb-database-create" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.185645 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7514750-68e9-462c-b401-f80689f8f4df" containerName="mariadb-database-create" Dec 03 06:47:43 crc kubenswrapper[4818]: E1203 06:47:43.185662 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd891b60-48cc-49f1-bb61-9e58c712816e" containerName="horizon" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.185667 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd891b60-48cc-49f1-bb61-9e58c712816e" containerName="horizon" Dec 03 06:47:43 crc kubenswrapper[4818]: E1203 06:47:43.185682 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" containerName="glance-log" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.185689 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" containerName="glance-log" Dec 03 06:47:43 crc kubenswrapper[4818]: E1203 06:47:43.185699 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae575e67-4d14-4f4c-8d27-b74f9b837134" containerName="mariadb-database-create" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.185705 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae575e67-4d14-4f4c-8d27-b74f9b837134" containerName="mariadb-database-create" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.185914 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" containerName="glance-httpd" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.185931 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd891b60-48cc-49f1-bb61-9e58c712816e" containerName="horizon-log" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.185946 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7514750-68e9-462c-b401-f80689f8f4df" containerName="mariadb-database-create" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.185959 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae575e67-4d14-4f4c-8d27-b74f9b837134" containerName="mariadb-database-create" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.185967 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2194aad-43ff-4ab6-b648-e079a0688a5f" containerName="mariadb-account-create-update" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.185981 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" containerName="glance-log" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.185991 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd891b60-48cc-49f1-bb61-9e58c712816e" containerName="horizon" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.187415 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.193230 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.193626 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.200855 4818 scope.go:117] "RemoveContainer" containerID="f5d9e63e248172bb123c2119799b48c42c47dc27abf1ccde1c22c8db92974276" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.202144 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae575e67-4d14-4f4c-8d27-b74f9b837134-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.202172 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7514750-68e9-462c-b401-f80689f8f4df-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.202182 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvc2n\" (UniqueName: \"kubernetes.io/projected/f2194aad-43ff-4ab6-b648-e079a0688a5f-kube-api-access-fvc2n\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.202190 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjq9l\" (UniqueName: \"kubernetes.io/projected/f7514750-68e9-462c-b401-f80689f8f4df-kube-api-access-pjq9l\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.202199 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlm9v\" (UniqueName: \"kubernetes.io/projected/ae575e67-4d14-4f4c-8d27-b74f9b837134-kube-api-access-wlm9v\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.202207 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2194aad-43ff-4ab6-b648-e079a0688a5f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.213476 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.304228 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22a6b677-2df1-41ed-a420-dbc5ce60afc5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.304264 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j9kt\" (UniqueName: \"kubernetes.io/projected/22a6b677-2df1-41ed-a420-dbc5ce60afc5-kube-api-access-2j9kt\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.304284 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.304315 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/22a6b677-2df1-41ed-a420-dbc5ce60afc5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.304337 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22a6b677-2df1-41ed-a420-dbc5ce60afc5-logs\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.304370 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22a6b677-2df1-41ed-a420-dbc5ce60afc5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.304395 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a6b677-2df1-41ed-a420-dbc5ce60afc5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.304461 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22a6b677-2df1-41ed-a420-dbc5ce60afc5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.405782 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/22a6b677-2df1-41ed-a420-dbc5ce60afc5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.406157 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22a6b677-2df1-41ed-a420-dbc5ce60afc5-logs\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.406195 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22a6b677-2df1-41ed-a420-dbc5ce60afc5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.406225 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a6b677-2df1-41ed-a420-dbc5ce60afc5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.406297 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22a6b677-2df1-41ed-a420-dbc5ce60afc5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.406351 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22a6b677-2df1-41ed-a420-dbc5ce60afc5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.406370 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j9kt\" (UniqueName: \"kubernetes.io/projected/22a6b677-2df1-41ed-a420-dbc5ce60afc5-kube-api-access-2j9kt\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.406388 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.406523 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.406761 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22a6b677-2df1-41ed-a420-dbc5ce60afc5-logs\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.406518 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/22a6b677-2df1-41ed-a420-dbc5ce60afc5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.414619 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22a6b677-2df1-41ed-a420-dbc5ce60afc5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.417063 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a6b677-2df1-41ed-a420-dbc5ce60afc5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.417837 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22a6b677-2df1-41ed-a420-dbc5ce60afc5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.418558 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22a6b677-2df1-41ed-a420-dbc5ce60afc5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.425768 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j9kt\" (UniqueName: \"kubernetes.io/projected/22a6b677-2df1-41ed-a420-dbc5ce60afc5-kube-api-access-2j9kt\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.471802 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"22a6b677-2df1-41ed-a420-dbc5ce60afc5\") " pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.518888 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.524803 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-sb9lv" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.743935 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdbn2\" (UniqueName: \"kubernetes.io/projected/5184be52-db4d-4e95-88bd-a6569ddb75ff-kube-api-access-zdbn2\") pod \"5184be52-db4d-4e95-88bd-a6569ddb75ff\" (UID: \"5184be52-db4d-4e95-88bd-a6569ddb75ff\") " Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.744119 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5184be52-db4d-4e95-88bd-a6569ddb75ff-operator-scripts\") pod \"5184be52-db4d-4e95-88bd-a6569ddb75ff\" (UID: \"5184be52-db4d-4e95-88bd-a6569ddb75ff\") " Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.746107 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5184be52-db4d-4e95-88bd-a6569ddb75ff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5184be52-db4d-4e95-88bd-a6569ddb75ff" (UID: "5184be52-db4d-4e95-88bd-a6569ddb75ff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.746613 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5184be52-db4d-4e95-88bd-a6569ddb75ff-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.787927 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5184be52-db4d-4e95-88bd-a6569ddb75ff-kube-api-access-zdbn2" (OuterVolumeSpecName: "kube-api-access-zdbn2") pod "5184be52-db4d-4e95-88bd-a6569ddb75ff" (UID: "5184be52-db4d-4e95-88bd-a6569ddb75ff"). InnerVolumeSpecName "kube-api-access-zdbn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.849510 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdbn2\" (UniqueName: \"kubernetes.io/projected/5184be52-db4d-4e95-88bd-a6569ddb75ff-kube-api-access-zdbn2\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:43 crc kubenswrapper[4818]: I1203 06:47:43.907854 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5631-account-create-update-k87bh" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.056007 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dcf3310-6518-4c25-a81c-ef30724db1b8-operator-scripts\") pod \"4dcf3310-6518-4c25-a81c-ef30724db1b8\" (UID: \"4dcf3310-6518-4c25-a81c-ef30724db1b8\") " Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.056391 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5bps\" (UniqueName: \"kubernetes.io/projected/4dcf3310-6518-4c25-a81c-ef30724db1b8-kube-api-access-g5bps\") pod \"4dcf3310-6518-4c25-a81c-ef30724db1b8\" (UID: \"4dcf3310-6518-4c25-a81c-ef30724db1b8\") " Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.056460 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dcf3310-6518-4c25-a81c-ef30724db1b8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4dcf3310-6518-4c25-a81c-ef30724db1b8" (UID: "4dcf3310-6518-4c25-a81c-ef30724db1b8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.056637 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8ffc-account-create-update-s65jt" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.056833 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dcf3310-6518-4c25-a81c-ef30724db1b8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.079573 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dcf3310-6518-4c25-a81c-ef30724db1b8-kube-api-access-g5bps" (OuterVolumeSpecName: "kube-api-access-g5bps") pod "4dcf3310-6518-4c25-a81c-ef30724db1b8" (UID: "4dcf3310-6518-4c25-a81c-ef30724db1b8"). InnerVolumeSpecName "kube-api-access-g5bps". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.130055 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5631-account-create-update-k87bh" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.130078 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5631-account-create-update-k87bh" event={"ID":"4dcf3310-6518-4c25-a81c-ef30724db1b8","Type":"ContainerDied","Data":"b60d29a02e8947cf9dc5db637aaacce93e351eec8db851fa064f755a2b15e6a6"} Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.130139 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b60d29a02e8947cf9dc5db637aaacce93e351eec8db851fa064f755a2b15e6a6" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.138745 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c46a9a7-6f86-40b8-8957-5d0f11d11262","Type":"ContainerStarted","Data":"3997bcfdd9b6ec0c28395631aa00ab5f718689df5dbdb47a3eab71a2ebce83aa"} Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.146354 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-sb9lv" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.149000 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-sb9lv" event={"ID":"5184be52-db4d-4e95-88bd-a6569ddb75ff","Type":"ContainerDied","Data":"9a5b024c89797b5d8e12e17458eacb212be6237dfb792e9eaab96ad014dd7a11"} Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.149047 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a5b024c89797b5d8e12e17458eacb212be6237dfb792e9eaab96ad014dd7a11" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.151545 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vn2ml" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.154002 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8ffc-account-create-update-s65jt" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.154179 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8ffc-account-create-update-s65jt" event={"ID":"7f0e2585-56c1-435c-9db7-74b19a1cab80","Type":"ContainerDied","Data":"48df0b580e2fa868b7da60dfc4768d0c1dbe9032758462364322868f06cbcd54"} Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.154212 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48df0b580e2fa868b7da60dfc4768d0c1dbe9032758462364322868f06cbcd54" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.154285 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-gff9c" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.160940 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f0e2585-56c1-435c-9db7-74b19a1cab80-operator-scripts\") pod \"7f0e2585-56c1-435c-9db7-74b19a1cab80\" (UID: \"7f0e2585-56c1-435c-9db7-74b19a1cab80\") " Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.161071 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtvs7\" (UniqueName: \"kubernetes.io/projected/7f0e2585-56c1-435c-9db7-74b19a1cab80-kube-api-access-jtvs7\") pod \"7f0e2585-56c1-435c-9db7-74b19a1cab80\" (UID: \"7f0e2585-56c1-435c-9db7-74b19a1cab80\") " Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.161407 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f0e2585-56c1-435c-9db7-74b19a1cab80-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7f0e2585-56c1-435c-9db7-74b19a1cab80" (UID: "7f0e2585-56c1-435c-9db7-74b19a1cab80"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.162488 4818 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f0e2585-56c1-435c-9db7-74b19a1cab80-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.162511 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5bps\" (UniqueName: \"kubernetes.io/projected/4dcf3310-6518-4c25-a81c-ef30724db1b8-kube-api-access-g5bps\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.164155 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f0e2585-56c1-435c-9db7-74b19a1cab80-kube-api-access-jtvs7" (OuterVolumeSpecName: "kube-api-access-jtvs7") pod "7f0e2585-56c1-435c-9db7-74b19a1cab80" (UID: "7f0e2585-56c1-435c-9db7-74b19a1cab80"). InnerVolumeSpecName "kube-api-access-jtvs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.264178 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtvs7\" (UniqueName: \"kubernetes.io/projected/7f0e2585-56c1-435c-9db7-74b19a1cab80-kube-api-access-jtvs7\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.532734 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 06:47:44 crc kubenswrapper[4818]: W1203 06:47:44.535472 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22a6b677_2df1_41ed_a420_dbc5ce60afc5.slice/crio-fa2f96ffa2ae881cd1ee956d29e05b607033841ee0b85765bf493519aaef8d45 WatchSource:0}: Error finding container fa2f96ffa2ae881cd1ee956d29e05b607033841ee0b85765bf493519aaef8d45: Status 404 returned error can't find the container with id fa2f96ffa2ae881cd1ee956d29e05b607033841ee0b85765bf493519aaef8d45 Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.781684 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11dff550-2d3d-4bbc-bbf8-e1bb6849c94f" path="/var/lib/kubelet/pods/11dff550-2d3d-4bbc-bbf8-e1bb6849c94f/volumes" Dec 03 06:47:44 crc kubenswrapper[4818]: I1203 06:47:44.782980 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd891b60-48cc-49f1-bb61-9e58c712816e" path="/var/lib/kubelet/pods/bd891b60-48cc-49f1-bb61-9e58c712816e/volumes" Dec 03 06:47:45 crc kubenswrapper[4818]: I1203 06:47:45.242570 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84520fa4-ba04-4475-ab4f-ab20f9ddc870","Type":"ContainerStarted","Data":"d064075747ffb7ae4d5cd6d3a0ca5fa0a4989c9881c388d391953271fa41f187"} Dec 03 06:47:45 crc kubenswrapper[4818]: I1203 06:47:45.242596 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerName="ceilometer-central-agent" containerID="cri-o://938292b21c559eca7770c1f64eae322064ebadab0d0155422f47e22e0dce950f" gracePeriod=30 Dec 03 06:47:45 crc kubenswrapper[4818]: I1203 06:47:45.242666 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerName="sg-core" containerID="cri-o://fc8f2925aa00371b84d68ec50ba5127addda07934af03dcc5db1cb124364777f" gracePeriod=30 Dec 03 06:47:45 crc kubenswrapper[4818]: I1203 06:47:45.242681 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerName="ceilometer-notification-agent" containerID="cri-o://dc4b5dcb248545aaf154ff7ee4db1271573719e5f4d9a5b34d9c617e48bb27bb" gracePeriod=30 Dec 03 06:47:45 crc kubenswrapper[4818]: I1203 06:47:45.242695 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 06:47:45 crc kubenswrapper[4818]: I1203 06:47:45.242684 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerName="proxy-httpd" containerID="cri-o://d064075747ffb7ae4d5cd6d3a0ca5fa0a4989c9881c388d391953271fa41f187" gracePeriod=30 Dec 03 06:47:45 crc kubenswrapper[4818]: I1203 06:47:45.249116 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"22a6b677-2df1-41ed-a420-dbc5ce60afc5","Type":"ContainerStarted","Data":"fa2f96ffa2ae881cd1ee956d29e05b607033841ee0b85765bf493519aaef8d45"} Dec 03 06:47:45 crc kubenswrapper[4818]: I1203 06:47:45.270695 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c46a9a7-6f86-40b8-8957-5d0f11d11262","Type":"ContainerStarted","Data":"f0a1b7418bd50718d817269f1bdf0860670de9d107de8bc5c58ddb880ac7da37"} Dec 03 06:47:45 crc kubenswrapper[4818]: I1203 06:47:45.272003 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.592920044 podStartE2EDuration="8.271983127s" podCreationTimestamp="2025-12-03 06:47:37 +0000 UTC" firstStartedPulling="2025-12-03 06:47:38.836290712 +0000 UTC m=+1216.527899464" lastFinishedPulling="2025-12-03 06:47:43.515353795 +0000 UTC m=+1221.206962547" observedRunningTime="2025-12-03 06:47:45.268530803 +0000 UTC m=+1222.960139555" watchObservedRunningTime="2025-12-03 06:47:45.271983127 +0000 UTC m=+1222.963591879" Dec 03 06:47:46 crc kubenswrapper[4818]: I1203 06:47:46.247140 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:46 crc kubenswrapper[4818]: I1203 06:47:46.265955 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7954d7ccff-th9lw" Dec 03 06:47:46 crc kubenswrapper[4818]: I1203 06:47:46.285426 4818 generic.go:334] "Generic (PLEG): container finished" podID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerID="d064075747ffb7ae4d5cd6d3a0ca5fa0a4989c9881c388d391953271fa41f187" exitCode=0 Dec 03 06:47:46 crc kubenswrapper[4818]: I1203 06:47:46.285458 4818 generic.go:334] "Generic (PLEG): container finished" podID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerID="fc8f2925aa00371b84d68ec50ba5127addda07934af03dcc5db1cb124364777f" exitCode=2 Dec 03 06:47:46 crc kubenswrapper[4818]: I1203 06:47:46.285468 4818 generic.go:334] "Generic (PLEG): container finished" podID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerID="dc4b5dcb248545aaf154ff7ee4db1271573719e5f4d9a5b34d9c617e48bb27bb" exitCode=0 Dec 03 06:47:46 crc kubenswrapper[4818]: I1203 06:47:46.285507 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84520fa4-ba04-4475-ab4f-ab20f9ddc870","Type":"ContainerDied","Data":"d064075747ffb7ae4d5cd6d3a0ca5fa0a4989c9881c388d391953271fa41f187"} Dec 03 06:47:46 crc kubenswrapper[4818]: I1203 06:47:46.285534 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84520fa4-ba04-4475-ab4f-ab20f9ddc870","Type":"ContainerDied","Data":"fc8f2925aa00371b84d68ec50ba5127addda07934af03dcc5db1cb124364777f"} Dec 03 06:47:46 crc kubenswrapper[4818]: I1203 06:47:46.285545 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84520fa4-ba04-4475-ab4f-ab20f9ddc870","Type":"ContainerDied","Data":"dc4b5dcb248545aaf154ff7ee4db1271573719e5f4d9a5b34d9c617e48bb27bb"} Dec 03 06:47:46 crc kubenswrapper[4818]: I1203 06:47:46.291320 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"22a6b677-2df1-41ed-a420-dbc5ce60afc5","Type":"ContainerStarted","Data":"ce059242b4c2c298d0e453de4eecdc984f948c41993f19bf52e961b641cb727b"} Dec 03 06:47:46 crc kubenswrapper[4818]: I1203 06:47:46.291377 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"22a6b677-2df1-41ed-a420-dbc5ce60afc5","Type":"ContainerStarted","Data":"d7baad49abda02c21dd82d59bb8fb52376be8bcb84a8bb0a3826e461ac05261b"} Dec 03 06:47:46 crc kubenswrapper[4818]: I1203 06:47:46.304557 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c46a9a7-6f86-40b8-8957-5d0f11d11262","Type":"ContainerStarted","Data":"c570a167a2676c4eea8ef810666e6cfe814f6a7ad6bc119e3a89e2b80104d3b2"} Dec 03 06:47:46 crc kubenswrapper[4818]: I1203 06:47:46.407050 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.407025694 podStartE2EDuration="3.407025694s" podCreationTimestamp="2025-12-03 06:47:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:47:46.361797404 +0000 UTC m=+1224.053406156" watchObservedRunningTime="2025-12-03 06:47:46.407025694 +0000 UTC m=+1224.098634446" Dec 03 06:47:46 crc kubenswrapper[4818]: I1203 06:47:46.442248 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.442228868 podStartE2EDuration="6.442228868s" podCreationTimestamp="2025-12-03 06:47:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:47:46.395672145 +0000 UTC m=+1224.087280897" watchObservedRunningTime="2025-12-03 06:47:46.442228868 +0000 UTC m=+1224.133837620" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.097292 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kxhw2"] Dec 03 06:47:50 crc kubenswrapper[4818]: E1203 06:47:50.097999 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f0e2585-56c1-435c-9db7-74b19a1cab80" containerName="mariadb-account-create-update" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.098018 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f0e2585-56c1-435c-9db7-74b19a1cab80" containerName="mariadb-account-create-update" Dec 03 06:47:50 crc kubenswrapper[4818]: E1203 06:47:50.098047 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5184be52-db4d-4e95-88bd-a6569ddb75ff" containerName="mariadb-database-create" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.098055 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="5184be52-db4d-4e95-88bd-a6569ddb75ff" containerName="mariadb-database-create" Dec 03 06:47:50 crc kubenswrapper[4818]: E1203 06:47:50.098080 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dcf3310-6518-4c25-a81c-ef30724db1b8" containerName="mariadb-account-create-update" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.098089 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dcf3310-6518-4c25-a81c-ef30724db1b8" containerName="mariadb-account-create-update" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.098279 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f0e2585-56c1-435c-9db7-74b19a1cab80" containerName="mariadb-account-create-update" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.098296 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="5184be52-db4d-4e95-88bd-a6569ddb75ff" containerName="mariadb-database-create" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.098314 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dcf3310-6518-4c25-a81c-ef30724db1b8" containerName="mariadb-account-create-update" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.099249 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kxhw2" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.101861 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2q5qs" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.102206 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.102466 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.114006 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kxhw2"] Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.259283 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-scripts\") pod \"nova-cell0-conductor-db-sync-kxhw2\" (UID: \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\") " pod="openstack/nova-cell0-conductor-db-sync-kxhw2" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.259636 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trd42\" (UniqueName: \"kubernetes.io/projected/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-kube-api-access-trd42\") pod \"nova-cell0-conductor-db-sync-kxhw2\" (UID: \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\") " pod="openstack/nova-cell0-conductor-db-sync-kxhw2" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.259766 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-config-data\") pod \"nova-cell0-conductor-db-sync-kxhw2\" (UID: \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\") " pod="openstack/nova-cell0-conductor-db-sync-kxhw2" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.259835 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kxhw2\" (UID: \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\") " pod="openstack/nova-cell0-conductor-db-sync-kxhw2" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.418365 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-scripts\") pod \"nova-cell0-conductor-db-sync-kxhw2\" (UID: \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\") " pod="openstack/nova-cell0-conductor-db-sync-kxhw2" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.418423 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trd42\" (UniqueName: \"kubernetes.io/projected/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-kube-api-access-trd42\") pod \"nova-cell0-conductor-db-sync-kxhw2\" (UID: \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\") " pod="openstack/nova-cell0-conductor-db-sync-kxhw2" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.418536 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-config-data\") pod \"nova-cell0-conductor-db-sync-kxhw2\" (UID: \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\") " pod="openstack/nova-cell0-conductor-db-sync-kxhw2" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.418578 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kxhw2\" (UID: \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\") " pod="openstack/nova-cell0-conductor-db-sync-kxhw2" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.426757 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kxhw2\" (UID: \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\") " pod="openstack/nova-cell0-conductor-db-sync-kxhw2" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.427558 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-config-data\") pod \"nova-cell0-conductor-db-sync-kxhw2\" (UID: \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\") " pod="openstack/nova-cell0-conductor-db-sync-kxhw2" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.431012 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-scripts\") pod \"nova-cell0-conductor-db-sync-kxhw2\" (UID: \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\") " pod="openstack/nova-cell0-conductor-db-sync-kxhw2" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.436196 4818 generic.go:334] "Generic (PLEG): container finished" podID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerID="938292b21c559eca7770c1f64eae322064ebadab0d0155422f47e22e0dce950f" exitCode=0 Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.436230 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84520fa4-ba04-4475-ab4f-ab20f9ddc870","Type":"ContainerDied","Data":"938292b21c559eca7770c1f64eae322064ebadab0d0155422f47e22e0dce950f"} Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.439788 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trd42\" (UniqueName: \"kubernetes.io/projected/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-kube-api-access-trd42\") pod \"nova-cell0-conductor-db-sync-kxhw2\" (UID: \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\") " pod="openstack/nova-cell0-conductor-db-sync-kxhw2" Dec 03 06:47:50 crc kubenswrapper[4818]: I1203 06:47:50.509061 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kxhw2" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.341798 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kxhw2"] Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.366618 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.366909 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.399138 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.417269 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.439309 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.446719 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kxhw2" event={"ID":"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987","Type":"ContainerStarted","Data":"d24b928b5e165ef50278dffaddc7ad1010bb58c702fc197e686b80a851e921b9"} Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.450273 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.451023 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"84520fa4-ba04-4475-ab4f-ab20f9ddc870","Type":"ContainerDied","Data":"445ca1e4e0c204f3c98228839f07fec9341ab2027b736636e9fdcb695560c9e9"} Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.451086 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.451109 4818 scope.go:117] "RemoveContainer" containerID="d064075747ffb7ae4d5cd6d3a0ca5fa0a4989c9881c388d391953271fa41f187" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.451776 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.481644 4818 scope.go:117] "RemoveContainer" containerID="fc8f2925aa00371b84d68ec50ba5127addda07934af03dcc5db1cb124364777f" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.524677 4818 scope.go:117] "RemoveContainer" containerID="dc4b5dcb248545aaf154ff7ee4db1271573719e5f4d9a5b34d9c617e48bb27bb" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.543922 4818 scope.go:117] "RemoveContainer" containerID="938292b21c559eca7770c1f64eae322064ebadab0d0155422f47e22e0dce950f" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.587379 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-config-data\") pod \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.587448 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4xbk\" (UniqueName: \"kubernetes.io/projected/84520fa4-ba04-4475-ab4f-ab20f9ddc870-kube-api-access-r4xbk\") pod \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.587508 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84520fa4-ba04-4475-ab4f-ab20f9ddc870-log-httpd\") pod \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.587555 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-sg-core-conf-yaml\") pod \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.587596 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-combined-ca-bundle\") pod \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.587643 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-scripts\") pod \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.587730 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84520fa4-ba04-4475-ab4f-ab20f9ddc870-run-httpd\") pod \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\" (UID: \"84520fa4-ba04-4475-ab4f-ab20f9ddc870\") " Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.588288 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84520fa4-ba04-4475-ab4f-ab20f9ddc870-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "84520fa4-ba04-4475-ab4f-ab20f9ddc870" (UID: "84520fa4-ba04-4475-ab4f-ab20f9ddc870"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.588379 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84520fa4-ba04-4475-ab4f-ab20f9ddc870-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "84520fa4-ba04-4475-ab4f-ab20f9ddc870" (UID: "84520fa4-ba04-4475-ab4f-ab20f9ddc870"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.597356 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84520fa4-ba04-4475-ab4f-ab20f9ddc870-kube-api-access-r4xbk" (OuterVolumeSpecName: "kube-api-access-r4xbk") pod "84520fa4-ba04-4475-ab4f-ab20f9ddc870" (UID: "84520fa4-ba04-4475-ab4f-ab20f9ddc870"). InnerVolumeSpecName "kube-api-access-r4xbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.598158 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-scripts" (OuterVolumeSpecName: "scripts") pod "84520fa4-ba04-4475-ab4f-ab20f9ddc870" (UID: "84520fa4-ba04-4475-ab4f-ab20f9ddc870"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.656958 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "84520fa4-ba04-4475-ab4f-ab20f9ddc870" (UID: "84520fa4-ba04-4475-ab4f-ab20f9ddc870"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.673697 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84520fa4-ba04-4475-ab4f-ab20f9ddc870" (UID: "84520fa4-ba04-4475-ab4f-ab20f9ddc870"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.777179 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4xbk\" (UniqueName: \"kubernetes.io/projected/84520fa4-ba04-4475-ab4f-ab20f9ddc870-kube-api-access-r4xbk\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.777621 4818 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84520fa4-ba04-4475-ab4f-ab20f9ddc870-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.778107 4818 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.778131 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.778141 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.778150 4818 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/84520fa4-ba04-4475-ab4f-ab20f9ddc870-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.896168 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-config-data" (OuterVolumeSpecName: "config-data") pod "84520fa4-ba04-4475-ab4f-ab20f9ddc870" (UID: "84520fa4-ba04-4475-ab4f-ab20f9ddc870"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:47:51 crc kubenswrapper[4818]: I1203 06:47:51.981696 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84520fa4-ba04-4475-ab4f-ab20f9ddc870-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.092786 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.105563 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.138197 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:52 crc kubenswrapper[4818]: E1203 06:47:52.138790 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerName="ceilometer-notification-agent" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.138828 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerName="ceilometer-notification-agent" Dec 03 06:47:52 crc kubenswrapper[4818]: E1203 06:47:52.138848 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerName="proxy-httpd" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.138856 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerName="proxy-httpd" Dec 03 06:47:52 crc kubenswrapper[4818]: E1203 06:47:52.138883 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerName="ceilometer-central-agent" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.138891 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerName="ceilometer-central-agent" Dec 03 06:47:52 crc kubenswrapper[4818]: E1203 06:47:52.138903 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerName="sg-core" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.138910 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerName="sg-core" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.139123 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerName="sg-core" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.139151 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerName="ceilometer-notification-agent" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.139164 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerName="proxy-httpd" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.139179 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" containerName="ceilometer-central-agent" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.141130 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.326049 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.326276 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.362100 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:52 crc kubenswrapper[4818]: E1203 06:47:52.362256 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84520fa4_ba04_4475_ab4f_ab20f9ddc870.slice/crio-445ca1e4e0c204f3c98228839f07fec9341ab2027b736636e9fdcb695560c9e9\": RecentStats: unable to find data in memory cache]" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.429170 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-config-data\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.429401 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-scripts\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.429428 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.429498 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.429525 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq7z4\" (UniqueName: \"kubernetes.io/projected/52deb96f-788b-4d52-a19f-c776f85c5903-kube-api-access-wq7z4\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.429552 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/52deb96f-788b-4d52-a19f-c776f85c5903-run-httpd\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.429574 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/52deb96f-788b-4d52-a19f-c776f85c5903-log-httpd\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.531065 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.531190 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.531232 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq7z4\" (UniqueName: \"kubernetes.io/projected/52deb96f-788b-4d52-a19f-c776f85c5903-kube-api-access-wq7z4\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.531277 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/52deb96f-788b-4d52-a19f-c776f85c5903-run-httpd\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.531308 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/52deb96f-788b-4d52-a19f-c776f85c5903-log-httpd\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.531371 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-config-data\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.531395 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-scripts\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.532870 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/52deb96f-788b-4d52-a19f-c776f85c5903-run-httpd\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.533734 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/52deb96f-788b-4d52-a19f-c776f85c5903-log-httpd\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.536235 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.537287 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-scripts\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.544846 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-config-data\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.545560 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.555977 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq7z4\" (UniqueName: \"kubernetes.io/projected/52deb96f-788b-4d52-a19f-c776f85c5903-kube-api-access-wq7z4\") pod \"ceilometer-0\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.723754 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:47:52 crc kubenswrapper[4818]: I1203 06:47:52.846148 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84520fa4-ba04-4475-ab4f-ab20f9ddc870" path="/var/lib/kubelet/pods/84520fa4-ba04-4475-ab4f-ab20f9ddc870/volumes" Dec 03 06:47:53 crc kubenswrapper[4818]: I1203 06:47:53.917196 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 06:47:53 crc kubenswrapper[4818]: I1203 06:47:53.917717 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 06:47:53 crc kubenswrapper[4818]: I1203 06:47:53.945205 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:47:53 crc kubenswrapper[4818]: I1203 06:47:53.945239 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:47:53 crc kubenswrapper[4818]: I1203 06:47:53.999092 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 06:47:53 crc kubenswrapper[4818]: I1203 06:47:53.999176 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 06:47:54 crc kubenswrapper[4818]: I1203 06:47:54.386338 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:54 crc kubenswrapper[4818]: W1203 06:47:54.393434 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52deb96f_788b_4d52_a19f_c776f85c5903.slice/crio-4ec47ce84bd354974be085fcc35ed66605a42641681e892b4cc4946bc23ba381 WatchSource:0}: Error finding container 4ec47ce84bd354974be085fcc35ed66605a42641681e892b4cc4946bc23ba381: Status 404 returned error can't find the container with id 4ec47ce84bd354974be085fcc35ed66605a42641681e892b4cc4946bc23ba381 Dec 03 06:47:54 crc kubenswrapper[4818]: I1203 06:47:54.974462 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"52deb96f-788b-4d52-a19f-c776f85c5903","Type":"ContainerStarted","Data":"4ec47ce84bd354974be085fcc35ed66605a42641681e892b4cc4946bc23ba381"} Dec 03 06:47:54 crc kubenswrapper[4818]: I1203 06:47:54.974939 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 06:47:54 crc kubenswrapper[4818]: I1203 06:47:54.974960 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 06:47:55 crc kubenswrapper[4818]: I1203 06:47:55.423614 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 06:47:55 crc kubenswrapper[4818]: I1203 06:47:55.424048 4818 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 06:47:55 crc kubenswrapper[4818]: I1203 06:47:55.751956 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 06:47:55 crc kubenswrapper[4818]: I1203 06:47:55.769574 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:47:56 crc kubenswrapper[4818]: I1203 06:47:56.148607 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"52deb96f-788b-4d52-a19f-c776f85c5903","Type":"ContainerStarted","Data":"0e04754da1e496e7c938da583bf89b8fa962542584455fda6010b9258cf2445c"} Dec 03 06:47:57 crc kubenswrapper[4818]: I1203 06:47:57.373217 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"52deb96f-788b-4d52-a19f-c776f85c5903","Type":"ContainerStarted","Data":"18072795b1e94a358165c236ac7b16d7dcb1de0b6620fef1b6b9ad604676f1e1"} Dec 03 06:47:58 crc kubenswrapper[4818]: I1203 06:47:58.551325 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"52deb96f-788b-4d52-a19f-c776f85c5903","Type":"ContainerStarted","Data":"b4a8c492b27a293b7efb8a6738e0bfc531138e29f963dcd86d7f1d5a6f0f731b"} Dec 03 06:47:59 crc kubenswrapper[4818]: I1203 06:47:59.581112 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 06:47:59 crc kubenswrapper[4818]: I1203 06:47:59.581210 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 06:48:00 crc kubenswrapper[4818]: I1203 06:48:00.715932 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"52deb96f-788b-4d52-a19f-c776f85c5903","Type":"ContainerStarted","Data":"e649ef27a897750be602887546aa8b0dab1d4aeadbfc942c3cf47bc8f566b318"} Dec 03 06:48:00 crc kubenswrapper[4818]: I1203 06:48:00.717758 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="52deb96f-788b-4d52-a19f-c776f85c5903" containerName="ceilometer-central-agent" containerID="cri-o://0e04754da1e496e7c938da583bf89b8fa962542584455fda6010b9258cf2445c" gracePeriod=30 Dec 03 06:48:00 crc kubenswrapper[4818]: I1203 06:48:00.717887 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 06:48:00 crc kubenswrapper[4818]: I1203 06:48:00.717916 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="52deb96f-788b-4d52-a19f-c776f85c5903" containerName="proxy-httpd" containerID="cri-o://e649ef27a897750be602887546aa8b0dab1d4aeadbfc942c3cf47bc8f566b318" gracePeriod=30 Dec 03 06:48:00 crc kubenswrapper[4818]: I1203 06:48:00.717958 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="52deb96f-788b-4d52-a19f-c776f85c5903" containerName="sg-core" containerID="cri-o://b4a8c492b27a293b7efb8a6738e0bfc531138e29f963dcd86d7f1d5a6f0f731b" gracePeriod=30 Dec 03 06:48:00 crc kubenswrapper[4818]: I1203 06:48:00.717989 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="52deb96f-788b-4d52-a19f-c776f85c5903" containerName="ceilometer-notification-agent" containerID="cri-o://18072795b1e94a358165c236ac7b16d7dcb1de0b6620fef1b6b9ad604676f1e1" gracePeriod=30 Dec 03 06:48:00 crc kubenswrapper[4818]: I1203 06:48:00.752659 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.448858331 podStartE2EDuration="8.752639837s" podCreationTimestamp="2025-12-03 06:47:52 +0000 UTC" firstStartedPulling="2025-12-03 06:47:54.400281385 +0000 UTC m=+1232.091890137" lastFinishedPulling="2025-12-03 06:47:59.704062891 +0000 UTC m=+1237.395671643" observedRunningTime="2025-12-03 06:48:00.743323698 +0000 UTC m=+1238.434932450" watchObservedRunningTime="2025-12-03 06:48:00.752639837 +0000 UTC m=+1238.444248599" Dec 03 06:48:01 crc kubenswrapper[4818]: I1203 06:48:01.734991 4818 generic.go:334] "Generic (PLEG): container finished" podID="52deb96f-788b-4d52-a19f-c776f85c5903" containerID="e649ef27a897750be602887546aa8b0dab1d4aeadbfc942c3cf47bc8f566b318" exitCode=0 Dec 03 06:48:01 crc kubenswrapper[4818]: I1203 06:48:01.735036 4818 generic.go:334] "Generic (PLEG): container finished" podID="52deb96f-788b-4d52-a19f-c776f85c5903" containerID="b4a8c492b27a293b7efb8a6738e0bfc531138e29f963dcd86d7f1d5a6f0f731b" exitCode=2 Dec 03 06:48:01 crc kubenswrapper[4818]: I1203 06:48:01.735049 4818 generic.go:334] "Generic (PLEG): container finished" podID="52deb96f-788b-4d52-a19f-c776f85c5903" containerID="18072795b1e94a358165c236ac7b16d7dcb1de0b6620fef1b6b9ad604676f1e1" exitCode=0 Dec 03 06:48:01 crc kubenswrapper[4818]: I1203 06:48:01.735076 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"52deb96f-788b-4d52-a19f-c776f85c5903","Type":"ContainerDied","Data":"e649ef27a897750be602887546aa8b0dab1d4aeadbfc942c3cf47bc8f566b318"} Dec 03 06:48:01 crc kubenswrapper[4818]: I1203 06:48:01.735111 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"52deb96f-788b-4d52-a19f-c776f85c5903","Type":"ContainerDied","Data":"b4a8c492b27a293b7efb8a6738e0bfc531138e29f963dcd86d7f1d5a6f0f731b"} Dec 03 06:48:01 crc kubenswrapper[4818]: I1203 06:48:01.735125 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"52deb96f-788b-4d52-a19f-c776f85c5903","Type":"ContainerDied","Data":"18072795b1e94a358165c236ac7b16d7dcb1de0b6620fef1b6b9ad604676f1e1"} Dec 03 06:48:08 crc kubenswrapper[4818]: I1203 06:48:08.953241 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kxhw2" event={"ID":"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987","Type":"ContainerStarted","Data":"b774aa954ffb4a76150f227eb0d04274bf155df99dc072805eac3f69b14c3b2f"} Dec 03 06:48:08 crc kubenswrapper[4818]: I1203 06:48:08.972492 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-kxhw2" podStartSLOduration=2.309022582 podStartE2EDuration="18.972476308s" podCreationTimestamp="2025-12-03 06:47:50 +0000 UTC" firstStartedPulling="2025-12-03 06:47:51.350003255 +0000 UTC m=+1229.041612017" lastFinishedPulling="2025-12-03 06:48:08.013456991 +0000 UTC m=+1245.705065743" observedRunningTime="2025-12-03 06:48:08.971936264 +0000 UTC m=+1246.663545036" watchObservedRunningTime="2025-12-03 06:48:08.972476308 +0000 UTC m=+1246.664085060" Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.609893 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.950870 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-scripts\") pod \"52deb96f-788b-4d52-a19f-c776f85c5903\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.950920 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/52deb96f-788b-4d52-a19f-c776f85c5903-log-httpd\") pod \"52deb96f-788b-4d52-a19f-c776f85c5903\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.951008 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-config-data\") pod \"52deb96f-788b-4d52-a19f-c776f85c5903\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.951050 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-sg-core-conf-yaml\") pod \"52deb96f-788b-4d52-a19f-c776f85c5903\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.951103 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-combined-ca-bundle\") pod \"52deb96f-788b-4d52-a19f-c776f85c5903\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.951131 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wq7z4\" (UniqueName: \"kubernetes.io/projected/52deb96f-788b-4d52-a19f-c776f85c5903-kube-api-access-wq7z4\") pod \"52deb96f-788b-4d52-a19f-c776f85c5903\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.951191 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/52deb96f-788b-4d52-a19f-c776f85c5903-run-httpd\") pod \"52deb96f-788b-4d52-a19f-c776f85c5903\" (UID: \"52deb96f-788b-4d52-a19f-c776f85c5903\") " Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.951923 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52deb96f-788b-4d52-a19f-c776f85c5903-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "52deb96f-788b-4d52-a19f-c776f85c5903" (UID: "52deb96f-788b-4d52-a19f-c776f85c5903"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.952472 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52deb96f-788b-4d52-a19f-c776f85c5903-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "52deb96f-788b-4d52-a19f-c776f85c5903" (UID: "52deb96f-788b-4d52-a19f-c776f85c5903"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.965831 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-scripts" (OuterVolumeSpecName: "scripts") pod "52deb96f-788b-4d52-a19f-c776f85c5903" (UID: "52deb96f-788b-4d52-a19f-c776f85c5903"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.966442 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52deb96f-788b-4d52-a19f-c776f85c5903-kube-api-access-wq7z4" (OuterVolumeSpecName: "kube-api-access-wq7z4") pod "52deb96f-788b-4d52-a19f-c776f85c5903" (UID: "52deb96f-788b-4d52-a19f-c776f85c5903"). InnerVolumeSpecName "kube-api-access-wq7z4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.974605 4818 generic.go:334] "Generic (PLEG): container finished" podID="52deb96f-788b-4d52-a19f-c776f85c5903" containerID="0e04754da1e496e7c938da583bf89b8fa962542584455fda6010b9258cf2445c" exitCode=0 Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.974859 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"52deb96f-788b-4d52-a19f-c776f85c5903","Type":"ContainerDied","Data":"0e04754da1e496e7c938da583bf89b8fa962542584455fda6010b9258cf2445c"} Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.974925 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"52deb96f-788b-4d52-a19f-c776f85c5903","Type":"ContainerDied","Data":"4ec47ce84bd354974be085fcc35ed66605a42641681e892b4cc4946bc23ba381"} Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.974950 4818 scope.go:117] "RemoveContainer" containerID="e649ef27a897750be602887546aa8b0dab1d4aeadbfc942c3cf47bc8f566b318" Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.975262 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:48:09 crc kubenswrapper[4818]: I1203 06:48:09.994072 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "52deb96f-788b-4d52-a19f-c776f85c5903" (UID: "52deb96f-788b-4d52-a19f-c776f85c5903"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.049914 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "52deb96f-788b-4d52-a19f-c776f85c5903" (UID: "52deb96f-788b-4d52-a19f-c776f85c5903"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.054018 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.054055 4818 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/52deb96f-788b-4d52-a19f-c776f85c5903-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.054073 4818 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.054091 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.054109 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wq7z4\" (UniqueName: \"kubernetes.io/projected/52deb96f-788b-4d52-a19f-c776f85c5903-kube-api-access-wq7z4\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.054123 4818 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/52deb96f-788b-4d52-a19f-c776f85c5903-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.072202 4818 scope.go:117] "RemoveContainer" containerID="b4a8c492b27a293b7efb8a6738e0bfc531138e29f963dcd86d7f1d5a6f0f731b" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.074933 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-config-data" (OuterVolumeSpecName: "config-data") pod "52deb96f-788b-4d52-a19f-c776f85c5903" (UID: "52deb96f-788b-4d52-a19f-c776f85c5903"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.098673 4818 scope.go:117] "RemoveContainer" containerID="18072795b1e94a358165c236ac7b16d7dcb1de0b6620fef1b6b9ad604676f1e1" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.128530 4818 scope.go:117] "RemoveContainer" containerID="0e04754da1e496e7c938da583bf89b8fa962542584455fda6010b9258cf2445c" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.150854 4818 scope.go:117] "RemoveContainer" containerID="e649ef27a897750be602887546aa8b0dab1d4aeadbfc942c3cf47bc8f566b318" Dec 03 06:48:10 crc kubenswrapper[4818]: E1203 06:48:10.151261 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e649ef27a897750be602887546aa8b0dab1d4aeadbfc942c3cf47bc8f566b318\": container with ID starting with e649ef27a897750be602887546aa8b0dab1d4aeadbfc942c3cf47bc8f566b318 not found: ID does not exist" containerID="e649ef27a897750be602887546aa8b0dab1d4aeadbfc942c3cf47bc8f566b318" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.151292 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e649ef27a897750be602887546aa8b0dab1d4aeadbfc942c3cf47bc8f566b318"} err="failed to get container status \"e649ef27a897750be602887546aa8b0dab1d4aeadbfc942c3cf47bc8f566b318\": rpc error: code = NotFound desc = could not find container \"e649ef27a897750be602887546aa8b0dab1d4aeadbfc942c3cf47bc8f566b318\": container with ID starting with e649ef27a897750be602887546aa8b0dab1d4aeadbfc942c3cf47bc8f566b318 not found: ID does not exist" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.151323 4818 scope.go:117] "RemoveContainer" containerID="b4a8c492b27a293b7efb8a6738e0bfc531138e29f963dcd86d7f1d5a6f0f731b" Dec 03 06:48:10 crc kubenswrapper[4818]: E1203 06:48:10.151562 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4a8c492b27a293b7efb8a6738e0bfc531138e29f963dcd86d7f1d5a6f0f731b\": container with ID starting with b4a8c492b27a293b7efb8a6738e0bfc531138e29f963dcd86d7f1d5a6f0f731b not found: ID does not exist" containerID="b4a8c492b27a293b7efb8a6738e0bfc531138e29f963dcd86d7f1d5a6f0f731b" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.151589 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4a8c492b27a293b7efb8a6738e0bfc531138e29f963dcd86d7f1d5a6f0f731b"} err="failed to get container status \"b4a8c492b27a293b7efb8a6738e0bfc531138e29f963dcd86d7f1d5a6f0f731b\": rpc error: code = NotFound desc = could not find container \"b4a8c492b27a293b7efb8a6738e0bfc531138e29f963dcd86d7f1d5a6f0f731b\": container with ID starting with b4a8c492b27a293b7efb8a6738e0bfc531138e29f963dcd86d7f1d5a6f0f731b not found: ID does not exist" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.151605 4818 scope.go:117] "RemoveContainer" containerID="18072795b1e94a358165c236ac7b16d7dcb1de0b6620fef1b6b9ad604676f1e1" Dec 03 06:48:10 crc kubenswrapper[4818]: E1203 06:48:10.151852 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18072795b1e94a358165c236ac7b16d7dcb1de0b6620fef1b6b9ad604676f1e1\": container with ID starting with 18072795b1e94a358165c236ac7b16d7dcb1de0b6620fef1b6b9ad604676f1e1 not found: ID does not exist" containerID="18072795b1e94a358165c236ac7b16d7dcb1de0b6620fef1b6b9ad604676f1e1" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.151874 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18072795b1e94a358165c236ac7b16d7dcb1de0b6620fef1b6b9ad604676f1e1"} err="failed to get container status \"18072795b1e94a358165c236ac7b16d7dcb1de0b6620fef1b6b9ad604676f1e1\": rpc error: code = NotFound desc = could not find container \"18072795b1e94a358165c236ac7b16d7dcb1de0b6620fef1b6b9ad604676f1e1\": container with ID starting with 18072795b1e94a358165c236ac7b16d7dcb1de0b6620fef1b6b9ad604676f1e1 not found: ID does not exist" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.151889 4818 scope.go:117] "RemoveContainer" containerID="0e04754da1e496e7c938da583bf89b8fa962542584455fda6010b9258cf2445c" Dec 03 06:48:10 crc kubenswrapper[4818]: E1203 06:48:10.152205 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e04754da1e496e7c938da583bf89b8fa962542584455fda6010b9258cf2445c\": container with ID starting with 0e04754da1e496e7c938da583bf89b8fa962542584455fda6010b9258cf2445c not found: ID does not exist" containerID="0e04754da1e496e7c938da583bf89b8fa962542584455fda6010b9258cf2445c" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.152226 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e04754da1e496e7c938da583bf89b8fa962542584455fda6010b9258cf2445c"} err="failed to get container status \"0e04754da1e496e7c938da583bf89b8fa962542584455fda6010b9258cf2445c\": rpc error: code = NotFound desc = could not find container \"0e04754da1e496e7c938da583bf89b8fa962542584455fda6010b9258cf2445c\": container with ID starting with 0e04754da1e496e7c938da583bf89b8fa962542584455fda6010b9258cf2445c not found: ID does not exist" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.155472 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52deb96f-788b-4d52-a19f-c776f85c5903-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.310655 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.321036 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.335754 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:48:10 crc kubenswrapper[4818]: E1203 06:48:10.336158 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52deb96f-788b-4d52-a19f-c776f85c5903" containerName="ceilometer-central-agent" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.336174 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="52deb96f-788b-4d52-a19f-c776f85c5903" containerName="ceilometer-central-agent" Dec 03 06:48:10 crc kubenswrapper[4818]: E1203 06:48:10.336186 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52deb96f-788b-4d52-a19f-c776f85c5903" containerName="proxy-httpd" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.336195 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="52deb96f-788b-4d52-a19f-c776f85c5903" containerName="proxy-httpd" Dec 03 06:48:10 crc kubenswrapper[4818]: E1203 06:48:10.336205 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52deb96f-788b-4d52-a19f-c776f85c5903" containerName="sg-core" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.336211 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="52deb96f-788b-4d52-a19f-c776f85c5903" containerName="sg-core" Dec 03 06:48:10 crc kubenswrapper[4818]: E1203 06:48:10.336231 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52deb96f-788b-4d52-a19f-c776f85c5903" containerName="ceilometer-notification-agent" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.336237 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="52deb96f-788b-4d52-a19f-c776f85c5903" containerName="ceilometer-notification-agent" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.336395 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="52deb96f-788b-4d52-a19f-c776f85c5903" containerName="ceilometer-notification-agent" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.336407 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="52deb96f-788b-4d52-a19f-c776f85c5903" containerName="sg-core" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.336422 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="52deb96f-788b-4d52-a19f-c776f85c5903" containerName="proxy-httpd" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.336431 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="52deb96f-788b-4d52-a19f-c776f85c5903" containerName="ceilometer-central-agent" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.338716 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.341556 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.349368 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.368340 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.498853 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-config-data\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.498936 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.499028 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f02e890-d82c-46f4-84ed-2e0b4c52c636-log-httpd\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.499086 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f02e890-d82c-46f4-84ed-2e0b4c52c636-run-httpd\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.499120 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2npb6\" (UniqueName: \"kubernetes.io/projected/9f02e890-d82c-46f4-84ed-2e0b4c52c636-kube-api-access-2npb6\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.499138 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-scripts\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.499304 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.602493 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f02e890-d82c-46f4-84ed-2e0b4c52c636-log-httpd\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.602554 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f02e890-d82c-46f4-84ed-2e0b4c52c636-run-httpd\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.602602 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2npb6\" (UniqueName: \"kubernetes.io/projected/9f02e890-d82c-46f4-84ed-2e0b4c52c636-kube-api-access-2npb6\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.602630 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-scripts\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.602666 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.602734 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-config-data\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.602779 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.603067 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f02e890-d82c-46f4-84ed-2e0b4c52c636-log-httpd\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.604011 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f02e890-d82c-46f4-84ed-2e0b4c52c636-run-httpd\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.607132 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-scripts\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.607137 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.608935 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.619759 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-config-data\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.620306 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2npb6\" (UniqueName: \"kubernetes.io/projected/9f02e890-d82c-46f4-84ed-2e0b4c52c636-kube-api-access-2npb6\") pod \"ceilometer-0\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.671254 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:48:10 crc kubenswrapper[4818]: I1203 06:48:10.748985 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52deb96f-788b-4d52-a19f-c776f85c5903" path="/var/lib/kubelet/pods/52deb96f-788b-4d52-a19f-c776f85c5903/volumes" Dec 03 06:48:11 crc kubenswrapper[4818]: W1203 06:48:11.172746 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f02e890_d82c_46f4_84ed_2e0b4c52c636.slice/crio-50d10ab51647dd92a5ddee5260a9091d90bc0efca046732225cfd2049c89bad8 WatchSource:0}: Error finding container 50d10ab51647dd92a5ddee5260a9091d90bc0efca046732225cfd2049c89bad8: Status 404 returned error can't find the container with id 50d10ab51647dd92a5ddee5260a9091d90bc0efca046732225cfd2049c89bad8 Dec 03 06:48:11 crc kubenswrapper[4818]: I1203 06:48:11.185794 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:48:11 crc kubenswrapper[4818]: I1203 06:48:11.998006 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f02e890-d82c-46f4-84ed-2e0b4c52c636","Type":"ContainerStarted","Data":"02e77da4f8089be95e03b39d914dabac0035ec22865290296e7050f6140ec5e3"} Dec 03 06:48:11 crc kubenswrapper[4818]: I1203 06:48:11.998272 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f02e890-d82c-46f4-84ed-2e0b4c52c636","Type":"ContainerStarted","Data":"50d10ab51647dd92a5ddee5260a9091d90bc0efca046732225cfd2049c89bad8"} Dec 03 06:48:13 crc kubenswrapper[4818]: I1203 06:48:13.010140 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f02e890-d82c-46f4-84ed-2e0b4c52c636","Type":"ContainerStarted","Data":"e1094db065d2b631c023e19d573ee1b7cc57cd74021d2bd67f17ce2ff080992d"} Dec 03 06:48:14 crc kubenswrapper[4818]: I1203 06:48:14.021946 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f02e890-d82c-46f4-84ed-2e0b4c52c636","Type":"ContainerStarted","Data":"ca949e3462c352721de59625046e82defe2c63692f631ae9f777ef322307a5d6"} Dec 03 06:48:15 crc kubenswrapper[4818]: I1203 06:48:15.035809 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f02e890-d82c-46f4-84ed-2e0b4c52c636","Type":"ContainerStarted","Data":"877978f070fb57bbbb477c01920bc68a116993b3e21aa7ce1478d83a3f019161"} Dec 03 06:48:15 crc kubenswrapper[4818]: I1203 06:48:15.036612 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 06:48:15 crc kubenswrapper[4818]: I1203 06:48:15.071072 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8810905409999998 podStartE2EDuration="5.07105371s" podCreationTimestamp="2025-12-03 06:48:10 +0000 UTC" firstStartedPulling="2025-12-03 06:48:11.177422112 +0000 UTC m=+1248.869030884" lastFinishedPulling="2025-12-03 06:48:14.367385301 +0000 UTC m=+1252.058994053" observedRunningTime="2025-12-03 06:48:15.063509725 +0000 UTC m=+1252.755118497" watchObservedRunningTime="2025-12-03 06:48:15.07105371 +0000 UTC m=+1252.762662472" Dec 03 06:48:19 crc kubenswrapper[4818]: I1203 06:48:19.074198 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:48:19 crc kubenswrapper[4818]: I1203 06:48:19.075892 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerName="ceilometer-central-agent" containerID="cri-o://02e77da4f8089be95e03b39d914dabac0035ec22865290296e7050f6140ec5e3" gracePeriod=30 Dec 03 06:48:19 crc kubenswrapper[4818]: I1203 06:48:19.076957 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerName="proxy-httpd" containerID="cri-o://877978f070fb57bbbb477c01920bc68a116993b3e21aa7ce1478d83a3f019161" gracePeriod=30 Dec 03 06:48:19 crc kubenswrapper[4818]: I1203 06:48:19.077046 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerName="ceilometer-notification-agent" containerID="cri-o://e1094db065d2b631c023e19d573ee1b7cc57cd74021d2bd67f17ce2ff080992d" gracePeriod=30 Dec 03 06:48:19 crc kubenswrapper[4818]: I1203 06:48:19.077096 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerName="sg-core" containerID="cri-o://ca949e3462c352721de59625046e82defe2c63692f631ae9f777ef322307a5d6" gracePeriod=30 Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.078193 4818 generic.go:334] "Generic (PLEG): container finished" podID="c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987" containerID="b774aa954ffb4a76150f227eb0d04274bf155df99dc072805eac3f69b14c3b2f" exitCode=0 Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.078372 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kxhw2" event={"ID":"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987","Type":"ContainerDied","Data":"b774aa954ffb4a76150f227eb0d04274bf155df99dc072805eac3f69b14c3b2f"} Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.095791 4818 generic.go:334] "Generic (PLEG): container finished" podID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerID="877978f070fb57bbbb477c01920bc68a116993b3e21aa7ce1478d83a3f019161" exitCode=0 Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.095837 4818 generic.go:334] "Generic (PLEG): container finished" podID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerID="ca949e3462c352721de59625046e82defe2c63692f631ae9f777ef322307a5d6" exitCode=2 Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.095846 4818 generic.go:334] "Generic (PLEG): container finished" podID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerID="e1094db065d2b631c023e19d573ee1b7cc57cd74021d2bd67f17ce2ff080992d" exitCode=0 Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.095867 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f02e890-d82c-46f4-84ed-2e0b4c52c636","Type":"ContainerDied","Data":"877978f070fb57bbbb477c01920bc68a116993b3e21aa7ce1478d83a3f019161"} Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.095892 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f02e890-d82c-46f4-84ed-2e0b4c52c636","Type":"ContainerDied","Data":"ca949e3462c352721de59625046e82defe2c63692f631ae9f777ef322307a5d6"} Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.095901 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f02e890-d82c-46f4-84ed-2e0b4c52c636","Type":"ContainerDied","Data":"e1094db065d2b631c023e19d573ee1b7cc57cd74021d2bd67f17ce2ff080992d"} Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.740197 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.791784 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-combined-ca-bundle\") pod \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.791986 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-config-data\") pod \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.792132 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-scripts\") pod \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.792173 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2npb6\" (UniqueName: \"kubernetes.io/projected/9f02e890-d82c-46f4-84ed-2e0b4c52c636-kube-api-access-2npb6\") pod \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.792200 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f02e890-d82c-46f4-84ed-2e0b4c52c636-run-httpd\") pod \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.792224 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f02e890-d82c-46f4-84ed-2e0b4c52c636-log-httpd\") pod \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.792244 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-sg-core-conf-yaml\") pod \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\" (UID: \"9f02e890-d82c-46f4-84ed-2e0b4c52c636\") " Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.793297 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f02e890-d82c-46f4-84ed-2e0b4c52c636-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9f02e890-d82c-46f4-84ed-2e0b4c52c636" (UID: "9f02e890-d82c-46f4-84ed-2e0b4c52c636"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.794032 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f02e890-d82c-46f4-84ed-2e0b4c52c636-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9f02e890-d82c-46f4-84ed-2e0b4c52c636" (UID: "9f02e890-d82c-46f4-84ed-2e0b4c52c636"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.797276 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-scripts" (OuterVolumeSpecName: "scripts") pod "9f02e890-d82c-46f4-84ed-2e0b4c52c636" (UID: "9f02e890-d82c-46f4-84ed-2e0b4c52c636"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.798405 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f02e890-d82c-46f4-84ed-2e0b4c52c636-kube-api-access-2npb6" (OuterVolumeSpecName: "kube-api-access-2npb6") pod "9f02e890-d82c-46f4-84ed-2e0b4c52c636" (UID: "9f02e890-d82c-46f4-84ed-2e0b4c52c636"). InnerVolumeSpecName "kube-api-access-2npb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.818811 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9f02e890-d82c-46f4-84ed-2e0b4c52c636" (UID: "9f02e890-d82c-46f4-84ed-2e0b4c52c636"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.862804 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f02e890-d82c-46f4-84ed-2e0b4c52c636" (UID: "9f02e890-d82c-46f4-84ed-2e0b4c52c636"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.890758 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-config-data" (OuterVolumeSpecName: "config-data") pod "9f02e890-d82c-46f4-84ed-2e0b4c52c636" (UID: "9f02e890-d82c-46f4-84ed-2e0b4c52c636"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.894680 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.894904 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2npb6\" (UniqueName: \"kubernetes.io/projected/9f02e890-d82c-46f4-84ed-2e0b4c52c636-kube-api-access-2npb6\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.895051 4818 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f02e890-d82c-46f4-84ed-2e0b4c52c636-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.895197 4818 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9f02e890-d82c-46f4-84ed-2e0b4c52c636-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.895314 4818 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.895445 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:20 crc kubenswrapper[4818]: I1203 06:48:20.895550 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f02e890-d82c-46f4-84ed-2e0b4c52c636-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.114556 4818 generic.go:334] "Generic (PLEG): container finished" podID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerID="02e77da4f8089be95e03b39d914dabac0035ec22865290296e7050f6140ec5e3" exitCode=0 Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.114666 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f02e890-d82c-46f4-84ed-2e0b4c52c636","Type":"ContainerDied","Data":"02e77da4f8089be95e03b39d914dabac0035ec22865290296e7050f6140ec5e3"} Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.115105 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9f02e890-d82c-46f4-84ed-2e0b4c52c636","Type":"ContainerDied","Data":"50d10ab51647dd92a5ddee5260a9091d90bc0efca046732225cfd2049c89bad8"} Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.115145 4818 scope.go:117] "RemoveContainer" containerID="877978f070fb57bbbb477c01920bc68a116993b3e21aa7ce1478d83a3f019161" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.114712 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.141717 4818 scope.go:117] "RemoveContainer" containerID="ca949e3462c352721de59625046e82defe2c63692f631ae9f777ef322307a5d6" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.160740 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.171033 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.188450 4818 scope.go:117] "RemoveContainer" containerID="e1094db065d2b631c023e19d573ee1b7cc57cd74021d2bd67f17ce2ff080992d" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.194671 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:48:21 crc kubenswrapper[4818]: E1203 06:48:21.195062 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerName="sg-core" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.195081 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerName="sg-core" Dec 03 06:48:21 crc kubenswrapper[4818]: E1203 06:48:21.195099 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerName="ceilometer-notification-agent" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.195107 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerName="ceilometer-notification-agent" Dec 03 06:48:21 crc kubenswrapper[4818]: E1203 06:48:21.195122 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerName="ceilometer-central-agent" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.195129 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerName="ceilometer-central-agent" Dec 03 06:48:21 crc kubenswrapper[4818]: E1203 06:48:21.195138 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerName="proxy-httpd" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.195145 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerName="proxy-httpd" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.195313 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerName="ceilometer-central-agent" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.195325 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerName="proxy-httpd" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.195338 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerName="sg-core" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.195345 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" containerName="ceilometer-notification-agent" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.197481 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.200681 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.201586 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.224831 4818 scope.go:117] "RemoveContainer" containerID="02e77da4f8089be95e03b39d914dabac0035ec22865290296e7050f6140ec5e3" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.248043 4818 scope.go:117] "RemoveContainer" containerID="877978f070fb57bbbb477c01920bc68a116993b3e21aa7ce1478d83a3f019161" Dec 03 06:48:21 crc kubenswrapper[4818]: E1203 06:48:21.248884 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"877978f070fb57bbbb477c01920bc68a116993b3e21aa7ce1478d83a3f019161\": container with ID starting with 877978f070fb57bbbb477c01920bc68a116993b3e21aa7ce1478d83a3f019161 not found: ID does not exist" containerID="877978f070fb57bbbb477c01920bc68a116993b3e21aa7ce1478d83a3f019161" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.248927 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"877978f070fb57bbbb477c01920bc68a116993b3e21aa7ce1478d83a3f019161"} err="failed to get container status \"877978f070fb57bbbb477c01920bc68a116993b3e21aa7ce1478d83a3f019161\": rpc error: code = NotFound desc = could not find container \"877978f070fb57bbbb477c01920bc68a116993b3e21aa7ce1478d83a3f019161\": container with ID starting with 877978f070fb57bbbb477c01920bc68a116993b3e21aa7ce1478d83a3f019161 not found: ID does not exist" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.248962 4818 scope.go:117] "RemoveContainer" containerID="ca949e3462c352721de59625046e82defe2c63692f631ae9f777ef322307a5d6" Dec 03 06:48:21 crc kubenswrapper[4818]: E1203 06:48:21.249254 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca949e3462c352721de59625046e82defe2c63692f631ae9f777ef322307a5d6\": container with ID starting with ca949e3462c352721de59625046e82defe2c63692f631ae9f777ef322307a5d6 not found: ID does not exist" containerID="ca949e3462c352721de59625046e82defe2c63692f631ae9f777ef322307a5d6" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.249280 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca949e3462c352721de59625046e82defe2c63692f631ae9f777ef322307a5d6"} err="failed to get container status \"ca949e3462c352721de59625046e82defe2c63692f631ae9f777ef322307a5d6\": rpc error: code = NotFound desc = could not find container \"ca949e3462c352721de59625046e82defe2c63692f631ae9f777ef322307a5d6\": container with ID starting with ca949e3462c352721de59625046e82defe2c63692f631ae9f777ef322307a5d6 not found: ID does not exist" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.249301 4818 scope.go:117] "RemoveContainer" containerID="e1094db065d2b631c023e19d573ee1b7cc57cd74021d2bd67f17ce2ff080992d" Dec 03 06:48:21 crc kubenswrapper[4818]: E1203 06:48:21.255359 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1094db065d2b631c023e19d573ee1b7cc57cd74021d2bd67f17ce2ff080992d\": container with ID starting with e1094db065d2b631c023e19d573ee1b7cc57cd74021d2bd67f17ce2ff080992d not found: ID does not exist" containerID="e1094db065d2b631c023e19d573ee1b7cc57cd74021d2bd67f17ce2ff080992d" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.255392 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1094db065d2b631c023e19d573ee1b7cc57cd74021d2bd67f17ce2ff080992d"} err="failed to get container status \"e1094db065d2b631c023e19d573ee1b7cc57cd74021d2bd67f17ce2ff080992d\": rpc error: code = NotFound desc = could not find container \"e1094db065d2b631c023e19d573ee1b7cc57cd74021d2bd67f17ce2ff080992d\": container with ID starting with e1094db065d2b631c023e19d573ee1b7cc57cd74021d2bd67f17ce2ff080992d not found: ID does not exist" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.255420 4818 scope.go:117] "RemoveContainer" containerID="02e77da4f8089be95e03b39d914dabac0035ec22865290296e7050f6140ec5e3" Dec 03 06:48:21 crc kubenswrapper[4818]: E1203 06:48:21.256079 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02e77da4f8089be95e03b39d914dabac0035ec22865290296e7050f6140ec5e3\": container with ID starting with 02e77da4f8089be95e03b39d914dabac0035ec22865290296e7050f6140ec5e3 not found: ID does not exist" containerID="02e77da4f8089be95e03b39d914dabac0035ec22865290296e7050f6140ec5e3" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.256103 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02e77da4f8089be95e03b39d914dabac0035ec22865290296e7050f6140ec5e3"} err="failed to get container status \"02e77da4f8089be95e03b39d914dabac0035ec22865290296e7050f6140ec5e3\": rpc error: code = NotFound desc = could not find container \"02e77da4f8089be95e03b39d914dabac0035ec22865290296e7050f6140ec5e3\": container with ID starting with 02e77da4f8089be95e03b39d914dabac0035ec22865290296e7050f6140ec5e3 not found: ID does not exist" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.265414 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.305146 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.305198 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-scripts\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.305223 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7e28850-7e84-401f-8015-9ed869354e9b-log-httpd\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.305255 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7e28850-7e84-401f-8015-9ed869354e9b-run-httpd\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.305292 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.305319 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46bh7\" (UniqueName: \"kubernetes.io/projected/b7e28850-7e84-401f-8015-9ed869354e9b-kube-api-access-46bh7\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.305371 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-config-data\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.407725 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-config-data\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.408252 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.408531 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-scripts\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.408555 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7e28850-7e84-401f-8015-9ed869354e9b-log-httpd\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.408780 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7e28850-7e84-401f-8015-9ed869354e9b-run-httpd\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.408942 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.408972 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46bh7\" (UniqueName: \"kubernetes.io/projected/b7e28850-7e84-401f-8015-9ed869354e9b-kube-api-access-46bh7\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.409299 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7e28850-7e84-401f-8015-9ed869354e9b-run-httpd\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.410053 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7e28850-7e84-401f-8015-9ed869354e9b-log-httpd\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.415660 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.415795 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-scripts\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.415990 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.417193 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-config-data\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.425332 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46bh7\" (UniqueName: \"kubernetes.io/projected/b7e28850-7e84-401f-8015-9ed869354e9b-kube-api-access-46bh7\") pod \"ceilometer-0\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.523570 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.528446 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kxhw2" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.613378 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-config-data\") pod \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\" (UID: \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\") " Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.613860 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-combined-ca-bundle\") pod \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\" (UID: \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\") " Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.613942 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trd42\" (UniqueName: \"kubernetes.io/projected/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-kube-api-access-trd42\") pod \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\" (UID: \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\") " Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.614023 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-scripts\") pod \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\" (UID: \"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987\") " Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.618283 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-kube-api-access-trd42" (OuterVolumeSpecName: "kube-api-access-trd42") pod "c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987" (UID: "c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987"). InnerVolumeSpecName "kube-api-access-trd42". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.618722 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-scripts" (OuterVolumeSpecName: "scripts") pod "c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987" (UID: "c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.647599 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-config-data" (OuterVolumeSpecName: "config-data") pod "c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987" (UID: "c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.662007 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987" (UID: "c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.716261 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.716296 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.716306 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trd42\" (UniqueName: \"kubernetes.io/projected/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-kube-api-access-trd42\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.716316 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:21 crc kubenswrapper[4818]: I1203 06:48:21.991000 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:48:21 crc kubenswrapper[4818]: W1203 06:48:21.991555 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7e28850_7e84_401f_8015_9ed869354e9b.slice/crio-cf324af06b14813bb47fe0eb305607db0ce82a33e812c0a1d7627d14b99ab09d WatchSource:0}: Error finding container cf324af06b14813bb47fe0eb305607db0ce82a33e812c0a1d7627d14b99ab09d: Status 404 returned error can't find the container with id cf324af06b14813bb47fe0eb305607db0ce82a33e812c0a1d7627d14b99ab09d Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.128431 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7e28850-7e84-401f-8015-9ed869354e9b","Type":"ContainerStarted","Data":"cf324af06b14813bb47fe0eb305607db0ce82a33e812c0a1d7627d14b99ab09d"} Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.130168 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kxhw2" event={"ID":"c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987","Type":"ContainerDied","Data":"d24b928b5e165ef50278dffaddc7ad1010bb58c702fc197e686b80a851e921b9"} Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.130199 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d24b928b5e165ef50278dffaddc7ad1010bb58c702fc197e686b80a851e921b9" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.130253 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kxhw2" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.186995 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 06:48:22 crc kubenswrapper[4818]: E1203 06:48:22.187760 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987" containerName="nova-cell0-conductor-db-sync" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.187782 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987" containerName="nova-cell0-conductor-db-sync" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.188025 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987" containerName="nova-cell0-conductor-db-sync" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.188880 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.193350 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-2q5qs" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.193765 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.199199 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.328932 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40ddc54-764b-4718-aea1-ac8e81eccae7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f40ddc54-764b-4718-aea1-ac8e81eccae7\") " pod="openstack/nova-cell0-conductor-0" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.328975 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40ddc54-764b-4718-aea1-ac8e81eccae7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f40ddc54-764b-4718-aea1-ac8e81eccae7\") " pod="openstack/nova-cell0-conductor-0" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.329137 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvgkt\" (UniqueName: \"kubernetes.io/projected/f40ddc54-764b-4718-aea1-ac8e81eccae7-kube-api-access-nvgkt\") pod \"nova-cell0-conductor-0\" (UID: \"f40ddc54-764b-4718-aea1-ac8e81eccae7\") " pod="openstack/nova-cell0-conductor-0" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.430522 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvgkt\" (UniqueName: \"kubernetes.io/projected/f40ddc54-764b-4718-aea1-ac8e81eccae7-kube-api-access-nvgkt\") pod \"nova-cell0-conductor-0\" (UID: \"f40ddc54-764b-4718-aea1-ac8e81eccae7\") " pod="openstack/nova-cell0-conductor-0" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.430607 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40ddc54-764b-4718-aea1-ac8e81eccae7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f40ddc54-764b-4718-aea1-ac8e81eccae7\") " pod="openstack/nova-cell0-conductor-0" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.430634 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40ddc54-764b-4718-aea1-ac8e81eccae7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f40ddc54-764b-4718-aea1-ac8e81eccae7\") " pod="openstack/nova-cell0-conductor-0" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.435049 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40ddc54-764b-4718-aea1-ac8e81eccae7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f40ddc54-764b-4718-aea1-ac8e81eccae7\") " pod="openstack/nova-cell0-conductor-0" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.436643 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40ddc54-764b-4718-aea1-ac8e81eccae7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f40ddc54-764b-4718-aea1-ac8e81eccae7\") " pod="openstack/nova-cell0-conductor-0" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.458245 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvgkt\" (UniqueName: \"kubernetes.io/projected/f40ddc54-764b-4718-aea1-ac8e81eccae7-kube-api-access-nvgkt\") pod \"nova-cell0-conductor-0\" (UID: \"f40ddc54-764b-4718-aea1-ac8e81eccae7\") " pod="openstack/nova-cell0-conductor-0" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.519733 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.750499 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f02e890-d82c-46f4-84ed-2e0b4c52c636" path="/var/lib/kubelet/pods/9f02e890-d82c-46f4-84ed-2e0b4c52c636/volumes" Dec 03 06:48:22 crc kubenswrapper[4818]: I1203 06:48:22.994768 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 06:48:22 crc kubenswrapper[4818]: W1203 06:48:22.994938 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf40ddc54_764b_4718_aea1_ac8e81eccae7.slice/crio-d80b2b464bf99dbec1e34200a728f70a822127cf68b0a26f66d3dc3c053eee5a WatchSource:0}: Error finding container d80b2b464bf99dbec1e34200a728f70a822127cf68b0a26f66d3dc3c053eee5a: Status 404 returned error can't find the container with id d80b2b464bf99dbec1e34200a728f70a822127cf68b0a26f66d3dc3c053eee5a Dec 03 06:48:23 crc kubenswrapper[4818]: I1203 06:48:23.140145 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f40ddc54-764b-4718-aea1-ac8e81eccae7","Type":"ContainerStarted","Data":"d80b2b464bf99dbec1e34200a728f70a822127cf68b0a26f66d3dc3c053eee5a"} Dec 03 06:48:23 crc kubenswrapper[4818]: I1203 06:48:23.141613 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7e28850-7e84-401f-8015-9ed869354e9b","Type":"ContainerStarted","Data":"98611168c65599f4c5e49c807527d3ba3df3c70f32abcb9b7302085e6abcaa2e"} Dec 03 06:48:24 crc kubenswrapper[4818]: I1203 06:48:24.150715 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7e28850-7e84-401f-8015-9ed869354e9b","Type":"ContainerStarted","Data":"b91146b25fb33b64f4db2d2d737328df7fbf72a90e372c764906b87477f71b72"} Dec 03 06:48:24 crc kubenswrapper[4818]: I1203 06:48:24.151342 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7e28850-7e84-401f-8015-9ed869354e9b","Type":"ContainerStarted","Data":"dafe75ff8f40e67a8d5af1ac2a59f8acbfd88e249efd0dd2df3f942379dd4674"} Dec 03 06:48:24 crc kubenswrapper[4818]: I1203 06:48:24.152844 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f40ddc54-764b-4718-aea1-ac8e81eccae7","Type":"ContainerStarted","Data":"2f979176dadb596707e277dbdffaa55ab2b9f9b8d86f5e9a28f905715e50a261"} Dec 03 06:48:24 crc kubenswrapper[4818]: I1203 06:48:24.153945 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 03 06:48:24 crc kubenswrapper[4818]: I1203 06:48:24.171376 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.171362751 podStartE2EDuration="2.171362751s" podCreationTimestamp="2025-12-03 06:48:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:48:24.169918496 +0000 UTC m=+1261.861527248" watchObservedRunningTime="2025-12-03 06:48:24.171362751 +0000 UTC m=+1261.862971503" Dec 03 06:48:25 crc kubenswrapper[4818]: I1203 06:48:25.163079 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7e28850-7e84-401f-8015-9ed869354e9b","Type":"ContainerStarted","Data":"5f45bb410933aa6bc90857efe0045b36269cffab652eaec93fefdc045f2fb23b"} Dec 03 06:48:25 crc kubenswrapper[4818]: I1203 06:48:25.163658 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 06:48:25 crc kubenswrapper[4818]: I1203 06:48:25.192681 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.245282891 podStartE2EDuration="4.192655056s" podCreationTimestamp="2025-12-03 06:48:21 +0000 UTC" firstStartedPulling="2025-12-03 06:48:21.994774333 +0000 UTC m=+1259.686383085" lastFinishedPulling="2025-12-03 06:48:24.942146498 +0000 UTC m=+1262.633755250" observedRunningTime="2025-12-03 06:48:25.186973757 +0000 UTC m=+1262.878582509" watchObservedRunningTime="2025-12-03 06:48:25.192655056 +0000 UTC m=+1262.884263808" Dec 03 06:48:32 crc kubenswrapper[4818]: I1203 06:48:32.558333 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.108685 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-d8cch"] Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.111432 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-d8cch" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.115075 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.120218 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.125341 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-d8cch"] Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.253423 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg92s\" (UniqueName: \"kubernetes.io/projected/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-kube-api-access-zg92s\") pod \"nova-cell0-cell-mapping-d8cch\" (UID: \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\") " pod="openstack/nova-cell0-cell-mapping-d8cch" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.253495 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-config-data\") pod \"nova-cell0-cell-mapping-d8cch\" (UID: \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\") " pod="openstack/nova-cell0-cell-mapping-d8cch" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.253597 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-scripts\") pod \"nova-cell0-cell-mapping-d8cch\" (UID: \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\") " pod="openstack/nova-cell0-cell-mapping-d8cch" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.253634 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-d8cch\" (UID: \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\") " pod="openstack/nova-cell0-cell-mapping-d8cch" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.319459 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.322516 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.334718 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.355050 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg92s\" (UniqueName: \"kubernetes.io/projected/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-kube-api-access-zg92s\") pod \"nova-cell0-cell-mapping-d8cch\" (UID: \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\") " pod="openstack/nova-cell0-cell-mapping-d8cch" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.355090 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-config-data\") pod \"nova-cell0-cell-mapping-d8cch\" (UID: \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\") " pod="openstack/nova-cell0-cell-mapping-d8cch" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.355154 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-scripts\") pod \"nova-cell0-cell-mapping-d8cch\" (UID: \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\") " pod="openstack/nova-cell0-cell-mapping-d8cch" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.355185 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-d8cch\" (UID: \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\") " pod="openstack/nova-cell0-cell-mapping-d8cch" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.396150 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.420476 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-d8cch\" (UID: \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\") " pod="openstack/nova-cell0-cell-mapping-d8cch" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.437973 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-config-data\") pod \"nova-cell0-cell-mapping-d8cch\" (UID: \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\") " pod="openstack/nova-cell0-cell-mapping-d8cch" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.464995 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/713fc838-5ff8-4799-876e-e21d870abbc0-config-data\") pod \"nova-api-0\" (UID: \"713fc838-5ff8-4799-876e-e21d870abbc0\") " pod="openstack/nova-api-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.465163 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/713fc838-5ff8-4799-876e-e21d870abbc0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"713fc838-5ff8-4799-876e-e21d870abbc0\") " pod="openstack/nova-api-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.465201 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/713fc838-5ff8-4799-876e-e21d870abbc0-logs\") pod \"nova-api-0\" (UID: \"713fc838-5ff8-4799-876e-e21d870abbc0\") " pod="openstack/nova-api-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.465331 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpxk2\" (UniqueName: \"kubernetes.io/projected/713fc838-5ff8-4799-876e-e21d870abbc0-kube-api-access-lpxk2\") pod \"nova-api-0\" (UID: \"713fc838-5ff8-4799-876e-e21d870abbc0\") " pod="openstack/nova-api-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.485314 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg92s\" (UniqueName: \"kubernetes.io/projected/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-kube-api-access-zg92s\") pod \"nova-cell0-cell-mapping-d8cch\" (UID: \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\") " pod="openstack/nova-cell0-cell-mapping-d8cch" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.516377 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-scripts\") pod \"nova-cell0-cell-mapping-d8cch\" (UID: \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\") " pod="openstack/nova-cell0-cell-mapping-d8cch" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.562084 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.573845 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/713fc838-5ff8-4799-876e-e21d870abbc0-config-data\") pod \"nova-api-0\" (UID: \"713fc838-5ff8-4799-876e-e21d870abbc0\") " pod="openstack/nova-api-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.573961 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/713fc838-5ff8-4799-876e-e21d870abbc0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"713fc838-5ff8-4799-876e-e21d870abbc0\") " pod="openstack/nova-api-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.574018 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/713fc838-5ff8-4799-876e-e21d870abbc0-logs\") pod \"nova-api-0\" (UID: \"713fc838-5ff8-4799-876e-e21d870abbc0\") " pod="openstack/nova-api-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.574196 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpxk2\" (UniqueName: \"kubernetes.io/projected/713fc838-5ff8-4799-876e-e21d870abbc0-kube-api-access-lpxk2\") pod \"nova-api-0\" (UID: \"713fc838-5ff8-4799-876e-e21d870abbc0\") " pod="openstack/nova-api-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.585527 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/713fc838-5ff8-4799-876e-e21d870abbc0-logs\") pod \"nova-api-0\" (UID: \"713fc838-5ff8-4799-876e-e21d870abbc0\") " pod="openstack/nova-api-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.586998 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/713fc838-5ff8-4799-876e-e21d870abbc0-config-data\") pod \"nova-api-0\" (UID: \"713fc838-5ff8-4799-876e-e21d870abbc0\") " pod="openstack/nova-api-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.590490 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.590542 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/713fc838-5ff8-4799-876e-e21d870abbc0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"713fc838-5ff8-4799-876e-e21d870abbc0\") " pod="openstack/nova-api-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.620981 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.650418 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.662740 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpxk2\" (UniqueName: \"kubernetes.io/projected/713fc838-5ff8-4799-876e-e21d870abbc0-kube-api-access-lpxk2\") pod \"nova-api-0\" (UID: \"713fc838-5ff8-4799-876e-e21d870abbc0\") " pod="openstack/nova-api-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.679290 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j7ks\" (UniqueName: \"kubernetes.io/projected/591d232a-3b0e-4043-821b-003ce40090a2-kube-api-access-5j7ks\") pod \"nova-metadata-0\" (UID: \"591d232a-3b0e-4043-821b-003ce40090a2\") " pod="openstack/nova-metadata-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.679333 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/591d232a-3b0e-4043-821b-003ce40090a2-config-data\") pod \"nova-metadata-0\" (UID: \"591d232a-3b0e-4043-821b-003ce40090a2\") " pod="openstack/nova-metadata-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.679425 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/591d232a-3b0e-4043-821b-003ce40090a2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"591d232a-3b0e-4043-821b-003ce40090a2\") " pod="openstack/nova-metadata-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.679452 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/591d232a-3b0e-4043-821b-003ce40090a2-logs\") pod \"nova-metadata-0\" (UID: \"591d232a-3b0e-4043-821b-003ce40090a2\") " pod="openstack/nova-metadata-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.709044 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.710281 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.717708 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.725789 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-tzstk"] Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.727854 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.734696 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-d8cch" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.738571 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.749218 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-tzstk"] Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.751606 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.759559 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.761255 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.766227 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.784267 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j7ks\" (UniqueName: \"kubernetes.io/projected/591d232a-3b0e-4043-821b-003ce40090a2-kube-api-access-5j7ks\") pod \"nova-metadata-0\" (UID: \"591d232a-3b0e-4043-821b-003ce40090a2\") " pod="openstack/nova-metadata-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.784317 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/591d232a-3b0e-4043-821b-003ce40090a2-config-data\") pod \"nova-metadata-0\" (UID: \"591d232a-3b0e-4043-821b-003ce40090a2\") " pod="openstack/nova-metadata-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.784454 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/591d232a-3b0e-4043-821b-003ce40090a2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"591d232a-3b0e-4043-821b-003ce40090a2\") " pod="openstack/nova-metadata-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.784496 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/591d232a-3b0e-4043-821b-003ce40090a2-logs\") pod \"nova-metadata-0\" (UID: \"591d232a-3b0e-4043-821b-003ce40090a2\") " pod="openstack/nova-metadata-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.784953 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/591d232a-3b0e-4043-821b-003ce40090a2-logs\") pod \"nova-metadata-0\" (UID: \"591d232a-3b0e-4043-821b-003ce40090a2\") " pod="openstack/nova-metadata-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.792168 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/591d232a-3b0e-4043-821b-003ce40090a2-config-data\") pod \"nova-metadata-0\" (UID: \"591d232a-3b0e-4043-821b-003ce40090a2\") " pod="openstack/nova-metadata-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.792266 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/591d232a-3b0e-4043-821b-003ce40090a2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"591d232a-3b0e-4043-821b-003ce40090a2\") " pod="openstack/nova-metadata-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.805906 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.806693 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j7ks\" (UniqueName: \"kubernetes.io/projected/591d232a-3b0e-4043-821b-003ce40090a2-kube-api-access-5j7ks\") pod \"nova-metadata-0\" (UID: \"591d232a-3b0e-4043-821b-003ce40090a2\") " pod="openstack/nova-metadata-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.886315 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.887211 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-config\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.887239 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33f783e-79d3-41be-93e7-c703833c95aa-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a33f783e-79d3-41be-93e7-c703833c95aa\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.887267 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nf7f\" (UniqueName: \"kubernetes.io/projected/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-kube-api-access-5nf7f\") pod \"nova-scheduler-0\" (UID: \"3678d8c7-48e3-4c65-9c05-48a5b96d0fed\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.887696 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.887762 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzh8n\" (UniqueName: \"kubernetes.io/projected/a33f783e-79d3-41be-93e7-c703833c95aa-kube-api-access-qzh8n\") pod \"nova-cell1-novncproxy-0\" (UID: \"a33f783e-79d3-41be-93e7-c703833c95aa\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.887784 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s6mg\" (UniqueName: \"kubernetes.io/projected/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-kube-api-access-2s6mg\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.887851 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-dns-svc\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.887905 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-config-data\") pod \"nova-scheduler-0\" (UID: \"3678d8c7-48e3-4c65-9c05-48a5b96d0fed\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.887923 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.887944 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33f783e-79d3-41be-93e7-c703833c95aa-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a33f783e-79d3-41be-93e7-c703833c95aa\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.888020 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3678d8c7-48e3-4c65-9c05-48a5b96d0fed\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.992436 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-config-data\") pod \"nova-scheduler-0\" (UID: \"3678d8c7-48e3-4c65-9c05-48a5b96d0fed\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.992487 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.992777 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33f783e-79d3-41be-93e7-c703833c95aa-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a33f783e-79d3-41be-93e7-c703833c95aa\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.992862 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3678d8c7-48e3-4c65-9c05-48a5b96d0fed\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.992942 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.992981 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-config\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.993003 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33f783e-79d3-41be-93e7-c703833c95aa-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a33f783e-79d3-41be-93e7-c703833c95aa\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.993027 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nf7f\" (UniqueName: \"kubernetes.io/projected/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-kube-api-access-5nf7f\") pod \"nova-scheduler-0\" (UID: \"3678d8c7-48e3-4c65-9c05-48a5b96d0fed\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.993048 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.993087 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzh8n\" (UniqueName: \"kubernetes.io/projected/a33f783e-79d3-41be-93e7-c703833c95aa-kube-api-access-qzh8n\") pod \"nova-cell1-novncproxy-0\" (UID: \"a33f783e-79d3-41be-93e7-c703833c95aa\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.993114 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s6mg\" (UniqueName: \"kubernetes.io/projected/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-kube-api-access-2s6mg\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:33 crc kubenswrapper[4818]: I1203 06:48:33.993160 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-dns-svc\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:33.994413 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-dns-svc\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.001880 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-config-data\") pod \"nova-scheduler-0\" (UID: \"3678d8c7-48e3-4c65-9c05-48a5b96d0fed\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.006409 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3678d8c7-48e3-4c65-9c05-48a5b96d0fed\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.006671 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.008347 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33f783e-79d3-41be-93e7-c703833c95aa-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a33f783e-79d3-41be-93e7-c703833c95aa\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.011725 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nf7f\" (UniqueName: \"kubernetes.io/projected/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-kube-api-access-5nf7f\") pod \"nova-scheduler-0\" (UID: \"3678d8c7-48e3-4c65-9c05-48a5b96d0fed\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.014138 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33f783e-79d3-41be-93e7-c703833c95aa-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a33f783e-79d3-41be-93e7-c703833c95aa\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.014739 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzh8n\" (UniqueName: \"kubernetes.io/projected/a33f783e-79d3-41be-93e7-c703833c95aa-kube-api-access-qzh8n\") pod \"nova-cell1-novncproxy-0\" (UID: \"a33f783e-79d3-41be-93e7-c703833c95aa\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.014750 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.025622 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-config\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.030491 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s6mg\" (UniqueName: \"kubernetes.io/projected/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-kube-api-access-2s6mg\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.031419 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.031441 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-tzstk\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.045360 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.053850 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.094647 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.305729 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-ql2nx"] Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.307453 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-ql2nx" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.310853 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-ql2nx\" (UID: \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\") " pod="openstack/nova-cell1-conductor-db-sync-ql2nx" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.310867 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.310983 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-scripts\") pod \"nova-cell1-conductor-db-sync-ql2nx\" (UID: \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\") " pod="openstack/nova-cell1-conductor-db-sync-ql2nx" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.311012 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g2l4\" (UniqueName: \"kubernetes.io/projected/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-kube-api-access-4g2l4\") pod \"nova-cell1-conductor-db-sync-ql2nx\" (UID: \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\") " pod="openstack/nova-cell1-conductor-db-sync-ql2nx" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.311075 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-config-data\") pod \"nova-cell1-conductor-db-sync-ql2nx\" (UID: \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\") " pod="openstack/nova-cell1-conductor-db-sync-ql2nx" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.311741 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.315319 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-ql2nx"] Dec 03 06:48:34 crc kubenswrapper[4818]: W1203 06:48:34.339191 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c1214a4_c9ac_42a3_93d9_ac9ac361c544.slice/crio-9924e388580ab8308171997b0c5188c037f99d31603a33ef8e85a44b6d4e6d36 WatchSource:0}: Error finding container 9924e388580ab8308171997b0c5188c037f99d31603a33ef8e85a44b6d4e6d36: Status 404 returned error can't find the container with id 9924e388580ab8308171997b0c5188c037f99d31603a33ef8e85a44b6d4e6d36 Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.341725 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-d8cch"] Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.399300 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.413008 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-ql2nx\" (UID: \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\") " pod="openstack/nova-cell1-conductor-db-sync-ql2nx" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.413063 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-scripts\") pod \"nova-cell1-conductor-db-sync-ql2nx\" (UID: \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\") " pod="openstack/nova-cell1-conductor-db-sync-ql2nx" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.413084 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g2l4\" (UniqueName: \"kubernetes.io/projected/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-kube-api-access-4g2l4\") pod \"nova-cell1-conductor-db-sync-ql2nx\" (UID: \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\") " pod="openstack/nova-cell1-conductor-db-sync-ql2nx" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.413134 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-config-data\") pod \"nova-cell1-conductor-db-sync-ql2nx\" (UID: \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\") " pod="openstack/nova-cell1-conductor-db-sync-ql2nx" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.421484 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-ql2nx\" (UID: \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\") " pod="openstack/nova-cell1-conductor-db-sync-ql2nx" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.424119 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-config-data\") pod \"nova-cell1-conductor-db-sync-ql2nx\" (UID: \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\") " pod="openstack/nova-cell1-conductor-db-sync-ql2nx" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.424572 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-scripts\") pod \"nova-cell1-conductor-db-sync-ql2nx\" (UID: \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\") " pod="openstack/nova-cell1-conductor-db-sync-ql2nx" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.438558 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g2l4\" (UniqueName: \"kubernetes.io/projected/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-kube-api-access-4g2l4\") pod \"nova-cell1-conductor-db-sync-ql2nx\" (UID: \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\") " pod="openstack/nova-cell1-conductor-db-sync-ql2nx" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.640430 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-ql2nx" Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.707661 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 06:48:34 crc kubenswrapper[4818]: W1203 06:48:34.715183 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3678d8c7_48e3_4c65_9c05_48a5b96d0fed.slice/crio-bdc4b5eedd52166388e241cac0f2c6d773c6d55e49825a58a2fa4111c5f5a89f WatchSource:0}: Error finding container bdc4b5eedd52166388e241cac0f2c6d773c6d55e49825a58a2fa4111c5f5a89f: Status 404 returned error can't find the container with id bdc4b5eedd52166388e241cac0f2c6d773c6d55e49825a58a2fa4111c5f5a89f Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.716410 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:48:34 crc kubenswrapper[4818]: W1203 06:48:34.890332 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4882a5a6_4fb7_4b00_b8bf_c28e116b5c16.slice/crio-507b2612e1aa77e754128452015e5d4264779d3ac2fd661f2c02f8f58ece9a41 WatchSource:0}: Error finding container 507b2612e1aa77e754128452015e5d4264779d3ac2fd661f2c02f8f58ece9a41: Status 404 returned error can't find the container with id 507b2612e1aa77e754128452015e5d4264779d3ac2fd661f2c02f8f58ece9a41 Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.923555 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 06:48:34 crc kubenswrapper[4818]: I1203 06:48:34.935555 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-tzstk"] Dec 03 06:48:35 crc kubenswrapper[4818]: I1203 06:48:35.176653 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-ql2nx"] Dec 03 06:48:35 crc kubenswrapper[4818]: W1203 06:48:35.182513 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod199309f9_ae9c_41cf_bd3a_dadf0f9c1c6c.slice/crio-0dee71f6eb21d45ab2d53cae1994e6197401d2d3b269a47bd71327cd1f4aedb1 WatchSource:0}: Error finding container 0dee71f6eb21d45ab2d53cae1994e6197401d2d3b269a47bd71327cd1f4aedb1: Status 404 returned error can't find the container with id 0dee71f6eb21d45ab2d53cae1994e6197401d2d3b269a47bd71327cd1f4aedb1 Dec 03 06:48:35 crc kubenswrapper[4818]: I1203 06:48:35.260302 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"713fc838-5ff8-4799-876e-e21d870abbc0","Type":"ContainerStarted","Data":"c853299d0559d2b77e6edf8151e3caf664494d358b3eafa370220b5893511a01"} Dec 03 06:48:35 crc kubenswrapper[4818]: I1203 06:48:35.263578 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-ql2nx" event={"ID":"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c","Type":"ContainerStarted","Data":"0dee71f6eb21d45ab2d53cae1994e6197401d2d3b269a47bd71327cd1f4aedb1"} Dec 03 06:48:35 crc kubenswrapper[4818]: I1203 06:48:35.265449 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a33f783e-79d3-41be-93e7-c703833c95aa","Type":"ContainerStarted","Data":"19c47281cf174949c1c92e7ba8a39ee3f204f81ae5442baaa826d596dbfd3ab1"} Dec 03 06:48:35 crc kubenswrapper[4818]: I1203 06:48:35.269918 4818 generic.go:334] "Generic (PLEG): container finished" podID="4882a5a6-4fb7-4b00-b8bf-c28e116b5c16" containerID="395e1d492ca211df82ce93315e1fac07e8eda09362461138bb52fcf64d81198e" exitCode=0 Dec 03 06:48:35 crc kubenswrapper[4818]: I1203 06:48:35.269982 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-tzstk" event={"ID":"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16","Type":"ContainerDied","Data":"395e1d492ca211df82ce93315e1fac07e8eda09362461138bb52fcf64d81198e"} Dec 03 06:48:35 crc kubenswrapper[4818]: I1203 06:48:35.270005 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-tzstk" event={"ID":"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16","Type":"ContainerStarted","Data":"507b2612e1aa77e754128452015e5d4264779d3ac2fd661f2c02f8f58ece9a41"} Dec 03 06:48:35 crc kubenswrapper[4818]: I1203 06:48:35.278164 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"591d232a-3b0e-4043-821b-003ce40090a2","Type":"ContainerStarted","Data":"eebd31a491cc17dc3e7ef28e2fda4c6e812f04b3d7265f639f2d6543817e0b20"} Dec 03 06:48:35 crc kubenswrapper[4818]: I1203 06:48:35.282771 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3678d8c7-48e3-4c65-9c05-48a5b96d0fed","Type":"ContainerStarted","Data":"bdc4b5eedd52166388e241cac0f2c6d773c6d55e49825a58a2fa4111c5f5a89f"} Dec 03 06:48:35 crc kubenswrapper[4818]: I1203 06:48:35.285336 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-d8cch" event={"ID":"9c1214a4-c9ac-42a3-93d9-ac9ac361c544","Type":"ContainerStarted","Data":"20091f6cfd086f85c0abaf606821fed5cbf8f11a0ce460a565af3f3e2f6f9f0d"} Dec 03 06:48:35 crc kubenswrapper[4818]: I1203 06:48:35.285391 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-d8cch" event={"ID":"9c1214a4-c9ac-42a3-93d9-ac9ac361c544","Type":"ContainerStarted","Data":"9924e388580ab8308171997b0c5188c037f99d31603a33ef8e85a44b6d4e6d36"} Dec 03 06:48:36 crc kubenswrapper[4818]: I1203 06:48:36.296848 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-ql2nx" event={"ID":"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c","Type":"ContainerStarted","Data":"3e0177bc27efb929e78580f92516e4c3ff4cc29a316b0119f792dc40a42fd1ae"} Dec 03 06:48:36 crc kubenswrapper[4818]: I1203 06:48:36.304199 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-tzstk" event={"ID":"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16","Type":"ContainerStarted","Data":"ed1281a3e82bb1946ebfd13ddb73195dbfa3219ba39b8229231fbecab8c9c184"} Dec 03 06:48:36 crc kubenswrapper[4818]: I1203 06:48:36.304400 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:36 crc kubenswrapper[4818]: I1203 06:48:36.320065 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-d8cch" podStartSLOduration=3.320045766 podStartE2EDuration="3.320045766s" podCreationTimestamp="2025-12-03 06:48:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:48:35.331227598 +0000 UTC m=+1273.022836350" watchObservedRunningTime="2025-12-03 06:48:36.320045766 +0000 UTC m=+1274.011654518" Dec 03 06:48:36 crc kubenswrapper[4818]: I1203 06:48:36.326183 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-ql2nx" podStartSLOduration=2.326163426 podStartE2EDuration="2.326163426s" podCreationTimestamp="2025-12-03 06:48:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:48:36.314191422 +0000 UTC m=+1274.005800174" watchObservedRunningTime="2025-12-03 06:48:36.326163426 +0000 UTC m=+1274.017772178" Dec 03 06:48:36 crc kubenswrapper[4818]: I1203 06:48:36.788038 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-tzstk" podStartSLOduration=3.78801098 podStartE2EDuration="3.78801098s" podCreationTimestamp="2025-12-03 06:48:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:48:36.346242329 +0000 UTC m=+1274.037851101" watchObservedRunningTime="2025-12-03 06:48:36.78801098 +0000 UTC m=+1274.479619732" Dec 03 06:48:36 crc kubenswrapper[4818]: I1203 06:48:36.796728 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 06:48:36 crc kubenswrapper[4818]: I1203 06:48:36.802783 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:48:39 crc kubenswrapper[4818]: I1203 06:48:39.334544 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="591d232a-3b0e-4043-821b-003ce40090a2" containerName="nova-metadata-log" containerID="cri-o://dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8" gracePeriod=30 Dec 03 06:48:39 crc kubenswrapper[4818]: I1203 06:48:39.334536 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"591d232a-3b0e-4043-821b-003ce40090a2","Type":"ContainerStarted","Data":"e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476"} Dec 03 06:48:39 crc kubenswrapper[4818]: I1203 06:48:39.335291 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"591d232a-3b0e-4043-821b-003ce40090a2","Type":"ContainerStarted","Data":"dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8"} Dec 03 06:48:39 crc kubenswrapper[4818]: I1203 06:48:39.334675 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="591d232a-3b0e-4043-821b-003ce40090a2" containerName="nova-metadata-metadata" containerID="cri-o://e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476" gracePeriod=30 Dec 03 06:48:39 crc kubenswrapper[4818]: I1203 06:48:39.338189 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3678d8c7-48e3-4c65-9c05-48a5b96d0fed","Type":"ContainerStarted","Data":"231700dfd026f94d1588e43d93e8a8a7dce63750aae9af62ec8862c58010836b"} Dec 03 06:48:39 crc kubenswrapper[4818]: I1203 06:48:39.354490 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"713fc838-5ff8-4799-876e-e21d870abbc0","Type":"ContainerStarted","Data":"1dac9236c1acb12ed2e9b466a699c14085a4a6115d3b106889e255449cd77aaf"} Dec 03 06:48:39 crc kubenswrapper[4818]: I1203 06:48:39.354547 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"713fc838-5ff8-4799-876e-e21d870abbc0","Type":"ContainerStarted","Data":"95ecafd87c822e414bd1fc06f5f75f9b48564c46cc8acbcc900d227f6c28519e"} Dec 03 06:48:39 crc kubenswrapper[4818]: I1203 06:48:39.359256 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a33f783e-79d3-41be-93e7-c703833c95aa","Type":"ContainerStarted","Data":"935aa46d2cd7f030179cdc34fbc5116fc6332f35e7dedac191ef26bf7889e304"} Dec 03 06:48:39 crc kubenswrapper[4818]: I1203 06:48:39.359414 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="a33f783e-79d3-41be-93e7-c703833c95aa" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://935aa46d2cd7f030179cdc34fbc5116fc6332f35e7dedac191ef26bf7889e304" gracePeriod=30 Dec 03 06:48:39 crc kubenswrapper[4818]: I1203 06:48:39.362358 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.982387149 podStartE2EDuration="6.36233482s" podCreationTimestamp="2025-12-03 06:48:33 +0000 UTC" firstStartedPulling="2025-12-03 06:48:34.795256424 +0000 UTC m=+1272.486865166" lastFinishedPulling="2025-12-03 06:48:38.175204085 +0000 UTC m=+1275.866812837" observedRunningTime="2025-12-03 06:48:39.358746522 +0000 UTC m=+1277.050355304" watchObservedRunningTime="2025-12-03 06:48:39.36233482 +0000 UTC m=+1277.053943612" Dec 03 06:48:39 crc kubenswrapper[4818]: I1203 06:48:39.387969 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.639485623 podStartE2EDuration="6.387947909s" podCreationTimestamp="2025-12-03 06:48:33 +0000 UTC" firstStartedPulling="2025-12-03 06:48:34.426281248 +0000 UTC m=+1272.117890000" lastFinishedPulling="2025-12-03 06:48:38.174743534 +0000 UTC m=+1275.866352286" observedRunningTime="2025-12-03 06:48:39.381396108 +0000 UTC m=+1277.073004870" watchObservedRunningTime="2025-12-03 06:48:39.387947909 +0000 UTC m=+1277.079556661" Dec 03 06:48:39 crc kubenswrapper[4818]: I1203 06:48:39.417506 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.966170041 podStartE2EDuration="6.417490314s" podCreationTimestamp="2025-12-03 06:48:33 +0000 UTC" firstStartedPulling="2025-12-03 06:48:34.726414714 +0000 UTC m=+1272.418023466" lastFinishedPulling="2025-12-03 06:48:38.177734987 +0000 UTC m=+1275.869343739" observedRunningTime="2025-12-03 06:48:39.408694008 +0000 UTC m=+1277.100302780" watchObservedRunningTime="2025-12-03 06:48:39.417490314 +0000 UTC m=+1277.109099066" Dec 03 06:48:39 crc kubenswrapper[4818]: I1203 06:48:39.432616 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.146653631 podStartE2EDuration="6.432596375s" podCreationTimestamp="2025-12-03 06:48:33 +0000 UTC" firstStartedPulling="2025-12-03 06:48:34.890221105 +0000 UTC m=+1272.581829857" lastFinishedPulling="2025-12-03 06:48:38.176163839 +0000 UTC m=+1275.867772601" observedRunningTime="2025-12-03 06:48:39.43201469 +0000 UTC m=+1277.123623442" watchObservedRunningTime="2025-12-03 06:48:39.432596375 +0000 UTC m=+1277.124205127" Dec 03 06:48:39 crc kubenswrapper[4818]: I1203 06:48:39.973979 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.068618 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/591d232a-3b0e-4043-821b-003ce40090a2-combined-ca-bundle\") pod \"591d232a-3b0e-4043-821b-003ce40090a2\" (UID: \"591d232a-3b0e-4043-821b-003ce40090a2\") " Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.068661 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/591d232a-3b0e-4043-821b-003ce40090a2-config-data\") pod \"591d232a-3b0e-4043-821b-003ce40090a2\" (UID: \"591d232a-3b0e-4043-821b-003ce40090a2\") " Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.068731 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5j7ks\" (UniqueName: \"kubernetes.io/projected/591d232a-3b0e-4043-821b-003ce40090a2-kube-api-access-5j7ks\") pod \"591d232a-3b0e-4043-821b-003ce40090a2\" (UID: \"591d232a-3b0e-4043-821b-003ce40090a2\") " Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.068765 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/591d232a-3b0e-4043-821b-003ce40090a2-logs\") pod \"591d232a-3b0e-4043-821b-003ce40090a2\" (UID: \"591d232a-3b0e-4043-821b-003ce40090a2\") " Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.069147 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/591d232a-3b0e-4043-821b-003ce40090a2-logs" (OuterVolumeSpecName: "logs") pod "591d232a-3b0e-4043-821b-003ce40090a2" (UID: "591d232a-3b0e-4043-821b-003ce40090a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.069280 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/591d232a-3b0e-4043-821b-003ce40090a2-logs\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.073697 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/591d232a-3b0e-4043-821b-003ce40090a2-kube-api-access-5j7ks" (OuterVolumeSpecName: "kube-api-access-5j7ks") pod "591d232a-3b0e-4043-821b-003ce40090a2" (UID: "591d232a-3b0e-4043-821b-003ce40090a2"). InnerVolumeSpecName "kube-api-access-5j7ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.104136 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/591d232a-3b0e-4043-821b-003ce40090a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "591d232a-3b0e-4043-821b-003ce40090a2" (UID: "591d232a-3b0e-4043-821b-003ce40090a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.105843 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/591d232a-3b0e-4043-821b-003ce40090a2-config-data" (OuterVolumeSpecName: "config-data") pod "591d232a-3b0e-4043-821b-003ce40090a2" (UID: "591d232a-3b0e-4043-821b-003ce40090a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.171455 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5j7ks\" (UniqueName: \"kubernetes.io/projected/591d232a-3b0e-4043-821b-003ce40090a2-kube-api-access-5j7ks\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.171790 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/591d232a-3b0e-4043-821b-003ce40090a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.171880 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/591d232a-3b0e-4043-821b-003ce40090a2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.372598 4818 generic.go:334] "Generic (PLEG): container finished" podID="591d232a-3b0e-4043-821b-003ce40090a2" containerID="e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476" exitCode=0 Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.372637 4818 generic.go:334] "Generic (PLEG): container finished" podID="591d232a-3b0e-4043-821b-003ce40090a2" containerID="dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8" exitCode=143 Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.372654 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.372722 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"591d232a-3b0e-4043-821b-003ce40090a2","Type":"ContainerDied","Data":"e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476"} Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.372754 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"591d232a-3b0e-4043-821b-003ce40090a2","Type":"ContainerDied","Data":"dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8"} Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.372767 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"591d232a-3b0e-4043-821b-003ce40090a2","Type":"ContainerDied","Data":"eebd31a491cc17dc3e7ef28e2fda4c6e812f04b3d7265f639f2d6543817e0b20"} Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.372785 4818 scope.go:117] "RemoveContainer" containerID="e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.412875 4818 scope.go:117] "RemoveContainer" containerID="dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.413052 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.432886 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.455745 4818 scope.go:117] "RemoveContainer" containerID="e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476" Dec 03 06:48:40 crc kubenswrapper[4818]: E1203 06:48:40.459959 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476\": container with ID starting with e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476 not found: ID does not exist" containerID="e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.460017 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476"} err="failed to get container status \"e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476\": rpc error: code = NotFound desc = could not find container \"e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476\": container with ID starting with e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476 not found: ID does not exist" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.460048 4818 scope.go:117] "RemoveContainer" containerID="dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8" Dec 03 06:48:40 crc kubenswrapper[4818]: E1203 06:48:40.461584 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8\": container with ID starting with dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8 not found: ID does not exist" containerID="dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.461641 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8"} err="failed to get container status \"dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8\": rpc error: code = NotFound desc = could not find container \"dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8\": container with ID starting with dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8 not found: ID does not exist" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.461667 4818 scope.go:117] "RemoveContainer" containerID="e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.462016 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476"} err="failed to get container status \"e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476\": rpc error: code = NotFound desc = could not find container \"e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476\": container with ID starting with e3ec17c4f0aef7c082d977529c5c0cddf2dc1a3b2845990176e8277ce0d93476 not found: ID does not exist" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.462054 4818 scope.go:117] "RemoveContainer" containerID="dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.462318 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8"} err="failed to get container status \"dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8\": rpc error: code = NotFound desc = could not find container \"dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8\": container with ID starting with dd4bf7943d67a6a6636487286890c702813b9c1315c3ad7e5ef7965a965d79e8 not found: ID does not exist" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.483874 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:48:40 crc kubenswrapper[4818]: E1203 06:48:40.485210 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="591d232a-3b0e-4043-821b-003ce40090a2" containerName="nova-metadata-log" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.485239 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="591d232a-3b0e-4043-821b-003ce40090a2" containerName="nova-metadata-log" Dec 03 06:48:40 crc kubenswrapper[4818]: E1203 06:48:40.485303 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="591d232a-3b0e-4043-821b-003ce40090a2" containerName="nova-metadata-metadata" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.485313 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="591d232a-3b0e-4043-821b-003ce40090a2" containerName="nova-metadata-metadata" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.485769 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="591d232a-3b0e-4043-821b-003ce40090a2" containerName="nova-metadata-log" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.485828 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="591d232a-3b0e-4043-821b-003ce40090a2" containerName="nova-metadata-metadata" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.502740 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.509048 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.509776 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.515667 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.580347 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbh6h\" (UniqueName: \"kubernetes.io/projected/7476f08a-fa54-403d-a383-09a4f27fccda-kube-api-access-bbh6h\") pod \"nova-metadata-0\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.580664 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.580775 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.580891 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-config-data\") pod \"nova-metadata-0\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.580976 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7476f08a-fa54-403d-a383-09a4f27fccda-logs\") pod \"nova-metadata-0\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.682388 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbh6h\" (UniqueName: \"kubernetes.io/projected/7476f08a-fa54-403d-a383-09a4f27fccda-kube-api-access-bbh6h\") pod \"nova-metadata-0\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.682503 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.682562 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.682597 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-config-data\") pod \"nova-metadata-0\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.682618 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7476f08a-fa54-403d-a383-09a4f27fccda-logs\") pod \"nova-metadata-0\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.683178 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7476f08a-fa54-403d-a383-09a4f27fccda-logs\") pod \"nova-metadata-0\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.686344 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.686850 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.694876 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-config-data\") pod \"nova-metadata-0\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.715932 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbh6h\" (UniqueName: \"kubernetes.io/projected/7476f08a-fa54-403d-a383-09a4f27fccda-kube-api-access-bbh6h\") pod \"nova-metadata-0\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " pod="openstack/nova-metadata-0" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.749609 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="591d232a-3b0e-4043-821b-003ce40090a2" path="/var/lib/kubelet/pods/591d232a-3b0e-4043-821b-003ce40090a2/volumes" Dec 03 06:48:40 crc kubenswrapper[4818]: I1203 06:48:40.834441 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 06:48:41 crc kubenswrapper[4818]: I1203 06:48:41.329053 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:48:41 crc kubenswrapper[4818]: I1203 06:48:41.401363 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7476f08a-fa54-403d-a383-09a4f27fccda","Type":"ContainerStarted","Data":"6965af681d16251fb13c0268cca14c6c480f5396105bd78b1a4e24b37f393fcd"} Dec 03 06:48:42 crc kubenswrapper[4818]: I1203 06:48:42.410838 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7476f08a-fa54-403d-a383-09a4f27fccda","Type":"ContainerStarted","Data":"1d72249a0a24853a97ba44f4d6265a62ec90b1a31fcce0dfc25a3ee745ffa6ba"} Dec 03 06:48:42 crc kubenswrapper[4818]: I1203 06:48:42.411352 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7476f08a-fa54-403d-a383-09a4f27fccda","Type":"ContainerStarted","Data":"5d5b24d073d0d874ef70cce146136a2725317c15b18cd5dbef5c344377a25ae6"} Dec 03 06:48:42 crc kubenswrapper[4818]: I1203 06:48:42.522922 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.522900377 podStartE2EDuration="2.522900377s" podCreationTimestamp="2025-12-03 06:48:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:48:42.513169268 +0000 UTC m=+1280.204778010" watchObservedRunningTime="2025-12-03 06:48:42.522900377 +0000 UTC m=+1280.214509129" Dec 03 06:48:43 crc kubenswrapper[4818]: I1203 06:48:43.420784 4818 generic.go:334] "Generic (PLEG): container finished" podID="9c1214a4-c9ac-42a3-93d9-ac9ac361c544" containerID="20091f6cfd086f85c0abaf606821fed5cbf8f11a0ce460a565af3f3e2f6f9f0d" exitCode=0 Dec 03 06:48:43 crc kubenswrapper[4818]: I1203 06:48:43.420867 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-d8cch" event={"ID":"9c1214a4-c9ac-42a3-93d9-ac9ac361c544","Type":"ContainerDied","Data":"20091f6cfd086f85c0abaf606821fed5cbf8f11a0ce460a565af3f3e2f6f9f0d"} Dec 03 06:48:43 crc kubenswrapper[4818]: I1203 06:48:43.422800 4818 generic.go:334] "Generic (PLEG): container finished" podID="199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c" containerID="3e0177bc27efb929e78580f92516e4c3ff4cc29a316b0119f792dc40a42fd1ae" exitCode=0 Dec 03 06:48:43 crc kubenswrapper[4818]: I1203 06:48:43.423738 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-ql2nx" event={"ID":"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c","Type":"ContainerDied","Data":"3e0177bc27efb929e78580f92516e4c3ff4cc29a316b0119f792dc40a42fd1ae"} Dec 03 06:48:43 crc kubenswrapper[4818]: I1203 06:48:43.755691 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 06:48:43 crc kubenswrapper[4818]: I1203 06:48:43.755779 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.046125 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.046487 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.054986 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.082141 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.095170 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.145250 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-qzppx"] Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.145533 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" podUID="c27270e3-defd-4d5b-abcc-2f717189dbd5" containerName="dnsmasq-dns" containerID="cri-o://963ea9b3f0e8ad84c7d038bccd3864c2c651117a41c9203d8cbbc239c00df7fe" gracePeriod=10 Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.453524 4818 generic.go:334] "Generic (PLEG): container finished" podID="c27270e3-defd-4d5b-abcc-2f717189dbd5" containerID="963ea9b3f0e8ad84c7d038bccd3864c2c651117a41c9203d8cbbc239c00df7fe" exitCode=0 Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.453745 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" event={"ID":"c27270e3-defd-4d5b-abcc-2f717189dbd5","Type":"ContainerDied","Data":"963ea9b3f0e8ad84c7d038bccd3864c2c651117a41c9203d8cbbc239c00df7fe"} Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.524431 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.658078 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.743868 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-dns-svc\") pod \"c27270e3-defd-4d5b-abcc-2f717189dbd5\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.743928 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnckx\" (UniqueName: \"kubernetes.io/projected/c27270e3-defd-4d5b-abcc-2f717189dbd5-kube-api-access-hnckx\") pod \"c27270e3-defd-4d5b-abcc-2f717189dbd5\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.743954 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-ovsdbserver-sb\") pod \"c27270e3-defd-4d5b-abcc-2f717189dbd5\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.743977 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-ovsdbserver-nb\") pod \"c27270e3-defd-4d5b-abcc-2f717189dbd5\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.744039 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-config\") pod \"c27270e3-defd-4d5b-abcc-2f717189dbd5\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.744055 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-dns-swift-storage-0\") pod \"c27270e3-defd-4d5b-abcc-2f717189dbd5\" (UID: \"c27270e3-defd-4d5b-abcc-2f717189dbd5\") " Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.753122 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c27270e3-defd-4d5b-abcc-2f717189dbd5-kube-api-access-hnckx" (OuterVolumeSpecName: "kube-api-access-hnckx") pod "c27270e3-defd-4d5b-abcc-2f717189dbd5" (UID: "c27270e3-defd-4d5b-abcc-2f717189dbd5"). InnerVolumeSpecName "kube-api-access-hnckx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.834587 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c27270e3-defd-4d5b-abcc-2f717189dbd5" (UID: "c27270e3-defd-4d5b-abcc-2f717189dbd5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.839091 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="713fc838-5ff8-4799-876e-e21d870abbc0" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.839168 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="713fc838-5ff8-4799-876e-e21d870abbc0" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.846039 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.846075 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnckx\" (UniqueName: \"kubernetes.io/projected/c27270e3-defd-4d5b-abcc-2f717189dbd5-kube-api-access-hnckx\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.891175 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-d8cch" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.900912 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-config" (OuterVolumeSpecName: "config") pod "c27270e3-defd-4d5b-abcc-2f717189dbd5" (UID: "c27270e3-defd-4d5b-abcc-2f717189dbd5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.925282 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c27270e3-defd-4d5b-abcc-2f717189dbd5" (UID: "c27270e3-defd-4d5b-abcc-2f717189dbd5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.931924 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c27270e3-defd-4d5b-abcc-2f717189dbd5" (UID: "c27270e3-defd-4d5b-abcc-2f717189dbd5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.934490 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c27270e3-defd-4d5b-abcc-2f717189dbd5" (UID: "c27270e3-defd-4d5b-abcc-2f717189dbd5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.947260 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-ql2nx" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.948412 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.948446 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.948457 4818 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:44 crc kubenswrapper[4818]: I1203 06:48:44.948470 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c27270e3-defd-4d5b-abcc-2f717189dbd5-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.050001 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-scripts\") pod \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\" (UID: \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\") " Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.050085 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-combined-ca-bundle\") pod \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\" (UID: \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\") " Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.050113 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-config-data\") pod \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\" (UID: \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\") " Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.050138 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-combined-ca-bundle\") pod \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\" (UID: \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\") " Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.050164 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-config-data\") pod \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\" (UID: \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\") " Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.050214 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4g2l4\" (UniqueName: \"kubernetes.io/projected/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-kube-api-access-4g2l4\") pod \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\" (UID: \"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c\") " Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.050283 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zg92s\" (UniqueName: \"kubernetes.io/projected/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-kube-api-access-zg92s\") pod \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\" (UID: \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\") " Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.050374 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-scripts\") pod \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\" (UID: \"9c1214a4-c9ac-42a3-93d9-ac9ac361c544\") " Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.053978 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-scripts" (OuterVolumeSpecName: "scripts") pod "9c1214a4-c9ac-42a3-93d9-ac9ac361c544" (UID: "9c1214a4-c9ac-42a3-93d9-ac9ac361c544"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.054842 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-kube-api-access-zg92s" (OuterVolumeSpecName: "kube-api-access-zg92s") pod "9c1214a4-c9ac-42a3-93d9-ac9ac361c544" (UID: "9c1214a4-c9ac-42a3-93d9-ac9ac361c544"). InnerVolumeSpecName "kube-api-access-zg92s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.056394 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-scripts" (OuterVolumeSpecName: "scripts") pod "199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c" (UID: "199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.056689 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-kube-api-access-4g2l4" (OuterVolumeSpecName: "kube-api-access-4g2l4") pod "199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c" (UID: "199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c"). InnerVolumeSpecName "kube-api-access-4g2l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.079907 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-config-data" (OuterVolumeSpecName: "config-data") pod "9c1214a4-c9ac-42a3-93d9-ac9ac361c544" (UID: "9c1214a4-c9ac-42a3-93d9-ac9ac361c544"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.080008 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c" (UID: "199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.083454 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c1214a4-c9ac-42a3-93d9-ac9ac361c544" (UID: "9c1214a4-c9ac-42a3-93d9-ac9ac361c544"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.091801 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-config-data" (OuterVolumeSpecName: "config-data") pod "199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c" (UID: "199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.152370 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.152406 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.152421 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.152433 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.152446 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.152456 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4g2l4\" (UniqueName: \"kubernetes.io/projected/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c-kube-api-access-4g2l4\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.152468 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zg92s\" (UniqueName: \"kubernetes.io/projected/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-kube-api-access-zg92s\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.152479 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c1214a4-c9ac-42a3-93d9-ac9ac361c544-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.468265 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-d8cch" event={"ID":"9c1214a4-c9ac-42a3-93d9-ac9ac361c544","Type":"ContainerDied","Data":"9924e388580ab8308171997b0c5188c037f99d31603a33ef8e85a44b6d4e6d36"} Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.468570 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9924e388580ab8308171997b0c5188c037f99d31603a33ef8e85a44b6d4e6d36" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.468621 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-d8cch" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.491801 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-ql2nx" event={"ID":"199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c","Type":"ContainerDied","Data":"0dee71f6eb21d45ab2d53cae1994e6197401d2d3b269a47bd71327cd1f4aedb1"} Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.491874 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0dee71f6eb21d45ab2d53cae1994e6197401d2d3b269a47bd71327cd1f4aedb1" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.491974 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-ql2nx" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.524029 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" event={"ID":"c27270e3-defd-4d5b-abcc-2f717189dbd5","Type":"ContainerDied","Data":"5afd38016a169387ac6fa5f4edab9bc61d23d03ea0e15433dd46acac97b5a6e5"} Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.524135 4818 scope.go:117] "RemoveContainer" containerID="963ea9b3f0e8ad84c7d038bccd3864c2c651117a41c9203d8cbbc239c00df7fe" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.537168 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-qzppx" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.582830 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 06:48:45 crc kubenswrapper[4818]: E1203 06:48:45.583338 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c27270e3-defd-4d5b-abcc-2f717189dbd5" containerName="dnsmasq-dns" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.583356 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c27270e3-defd-4d5b-abcc-2f717189dbd5" containerName="dnsmasq-dns" Dec 03 06:48:45 crc kubenswrapper[4818]: E1203 06:48:45.583374 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c1214a4-c9ac-42a3-93d9-ac9ac361c544" containerName="nova-manage" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.583381 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c1214a4-c9ac-42a3-93d9-ac9ac361c544" containerName="nova-manage" Dec 03 06:48:45 crc kubenswrapper[4818]: E1203 06:48:45.583401 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c27270e3-defd-4d5b-abcc-2f717189dbd5" containerName="init" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.583407 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c27270e3-defd-4d5b-abcc-2f717189dbd5" containerName="init" Dec 03 06:48:45 crc kubenswrapper[4818]: E1203 06:48:45.583430 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c" containerName="nova-cell1-conductor-db-sync" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.583436 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c" containerName="nova-cell1-conductor-db-sync" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.583611 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="c27270e3-defd-4d5b-abcc-2f717189dbd5" containerName="dnsmasq-dns" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.583623 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c" containerName="nova-cell1-conductor-db-sync" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.583639 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c1214a4-c9ac-42a3-93d9-ac9ac361c544" containerName="nova-manage" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.584355 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.591602 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.593535 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.593631 4818 scope.go:117] "RemoveContainer" containerID="29f6bb64fef338a65b9e3e325130da2d0822d016e8aaa5bd246a9dcaf965d0b7" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.667118 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.667313 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="713fc838-5ff8-4799-876e-e21d870abbc0" containerName="nova-api-log" containerID="cri-o://95ecafd87c822e414bd1fc06f5f75f9b48564c46cc8acbcc900d227f6c28519e" gracePeriod=30 Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.667682 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="713fc838-5ff8-4799-876e-e21d870abbc0" containerName="nova-api-api" containerID="cri-o://1dac9236c1acb12ed2e9b466a699c14085a4a6115d3b106889e255449cd77aaf" gracePeriod=30 Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.678955 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d772d6d8-be96-4256-84a5-e9f4713ad4ec-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d772d6d8-be96-4256-84a5-e9f4713ad4ec\") " pod="openstack/nova-cell1-conductor-0" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.679080 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km7sn\" (UniqueName: \"kubernetes.io/projected/d772d6d8-be96-4256-84a5-e9f4713ad4ec-kube-api-access-km7sn\") pod \"nova-cell1-conductor-0\" (UID: \"d772d6d8-be96-4256-84a5-e9f4713ad4ec\") " pod="openstack/nova-cell1-conductor-0" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.679122 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d772d6d8-be96-4256-84a5-e9f4713ad4ec-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d772d6d8-be96-4256-84a5-e9f4713ad4ec\") " pod="openstack/nova-cell1-conductor-0" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.702867 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-qzppx"] Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.723347 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.751235 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-qzppx"] Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.764875 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.765167 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7476f08a-fa54-403d-a383-09a4f27fccda" containerName="nova-metadata-log" containerID="cri-o://5d5b24d073d0d874ef70cce146136a2725317c15b18cd5dbef5c344377a25ae6" gracePeriod=30 Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.765609 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="7476f08a-fa54-403d-a383-09a4f27fccda" containerName="nova-metadata-metadata" containerID="cri-o://1d72249a0a24853a97ba44f4d6265a62ec90b1a31fcce0dfc25a3ee745ffa6ba" gracePeriod=30 Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.781928 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km7sn\" (UniqueName: \"kubernetes.io/projected/d772d6d8-be96-4256-84a5-e9f4713ad4ec-kube-api-access-km7sn\") pod \"nova-cell1-conductor-0\" (UID: \"d772d6d8-be96-4256-84a5-e9f4713ad4ec\") " pod="openstack/nova-cell1-conductor-0" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.782002 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d772d6d8-be96-4256-84a5-e9f4713ad4ec-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d772d6d8-be96-4256-84a5-e9f4713ad4ec\") " pod="openstack/nova-cell1-conductor-0" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.782083 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d772d6d8-be96-4256-84a5-e9f4713ad4ec-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d772d6d8-be96-4256-84a5-e9f4713ad4ec\") " pod="openstack/nova-cell1-conductor-0" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.789778 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d772d6d8-be96-4256-84a5-e9f4713ad4ec-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d772d6d8-be96-4256-84a5-e9f4713ad4ec\") " pod="openstack/nova-cell1-conductor-0" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.790117 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d772d6d8-be96-4256-84a5-e9f4713ad4ec-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d772d6d8-be96-4256-84a5-e9f4713ad4ec\") " pod="openstack/nova-cell1-conductor-0" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.840877 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.841251 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.852463 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km7sn\" (UniqueName: \"kubernetes.io/projected/d772d6d8-be96-4256-84a5-e9f4713ad4ec-kube-api-access-km7sn\") pod \"nova-cell1-conductor-0\" (UID: \"d772d6d8-be96-4256-84a5-e9f4713ad4ec\") " pod="openstack/nova-cell1-conductor-0" Dec 03 06:48:45 crc kubenswrapper[4818]: I1203 06:48:45.938573 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.451024 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.554591 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d772d6d8-be96-4256-84a5-e9f4713ad4ec","Type":"ContainerStarted","Data":"7092e9b6a6f411d629a238be11fb14b1c80ff37925f2882b208521a3278032ed"} Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.557795 4818 generic.go:334] "Generic (PLEG): container finished" podID="713fc838-5ff8-4799-876e-e21d870abbc0" containerID="95ecafd87c822e414bd1fc06f5f75f9b48564c46cc8acbcc900d227f6c28519e" exitCode=143 Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.557881 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"713fc838-5ff8-4799-876e-e21d870abbc0","Type":"ContainerDied","Data":"95ecafd87c822e414bd1fc06f5f75f9b48564c46cc8acbcc900d227f6c28519e"} Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.559454 4818 generic.go:334] "Generic (PLEG): container finished" podID="7476f08a-fa54-403d-a383-09a4f27fccda" containerID="1d72249a0a24853a97ba44f4d6265a62ec90b1a31fcce0dfc25a3ee745ffa6ba" exitCode=0 Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.559467 4818 generic.go:334] "Generic (PLEG): container finished" podID="7476f08a-fa54-403d-a383-09a4f27fccda" containerID="5d5b24d073d0d874ef70cce146136a2725317c15b18cd5dbef5c344377a25ae6" exitCode=143 Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.559603 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="3678d8c7-48e3-4c65-9c05-48a5b96d0fed" containerName="nova-scheduler-scheduler" containerID="cri-o://231700dfd026f94d1588e43d93e8a8a7dce63750aae9af62ec8862c58010836b" gracePeriod=30 Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.559891 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7476f08a-fa54-403d-a383-09a4f27fccda","Type":"ContainerDied","Data":"1d72249a0a24853a97ba44f4d6265a62ec90b1a31fcce0dfc25a3ee745ffa6ba"} Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.559911 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7476f08a-fa54-403d-a383-09a4f27fccda","Type":"ContainerDied","Data":"5d5b24d073d0d874ef70cce146136a2725317c15b18cd5dbef5c344377a25ae6"} Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.645492 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.746797 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c27270e3-defd-4d5b-abcc-2f717189dbd5" path="/var/lib/kubelet/pods/c27270e3-defd-4d5b-abcc-2f717189dbd5/volumes" Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.813515 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-combined-ca-bundle\") pod \"7476f08a-fa54-403d-a383-09a4f27fccda\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.813958 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7476f08a-fa54-403d-a383-09a4f27fccda-logs\") pod \"7476f08a-fa54-403d-a383-09a4f27fccda\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.814063 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-nova-metadata-tls-certs\") pod \"7476f08a-fa54-403d-a383-09a4f27fccda\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.814207 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-config-data\") pod \"7476f08a-fa54-403d-a383-09a4f27fccda\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.814338 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbh6h\" (UniqueName: \"kubernetes.io/projected/7476f08a-fa54-403d-a383-09a4f27fccda-kube-api-access-bbh6h\") pod \"7476f08a-fa54-403d-a383-09a4f27fccda\" (UID: \"7476f08a-fa54-403d-a383-09a4f27fccda\") " Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.814531 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7476f08a-fa54-403d-a383-09a4f27fccda-logs" (OuterVolumeSpecName: "logs") pod "7476f08a-fa54-403d-a383-09a4f27fccda" (UID: "7476f08a-fa54-403d-a383-09a4f27fccda"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.815145 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7476f08a-fa54-403d-a383-09a4f27fccda-logs\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.818586 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7476f08a-fa54-403d-a383-09a4f27fccda-kube-api-access-bbh6h" (OuterVolumeSpecName: "kube-api-access-bbh6h") pod "7476f08a-fa54-403d-a383-09a4f27fccda" (UID: "7476f08a-fa54-403d-a383-09a4f27fccda"). InnerVolumeSpecName "kube-api-access-bbh6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.844159 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-config-data" (OuterVolumeSpecName: "config-data") pod "7476f08a-fa54-403d-a383-09a4f27fccda" (UID: "7476f08a-fa54-403d-a383-09a4f27fccda"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.845218 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7476f08a-fa54-403d-a383-09a4f27fccda" (UID: "7476f08a-fa54-403d-a383-09a4f27fccda"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.870760 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "7476f08a-fa54-403d-a383-09a4f27fccda" (UID: "7476f08a-fa54-403d-a383-09a4f27fccda"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.916583 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.916619 4818 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.916630 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7476f08a-fa54-403d-a383-09a4f27fccda-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:46 crc kubenswrapper[4818]: I1203 06:48:46.916638 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbh6h\" (UniqueName: \"kubernetes.io/projected/7476f08a-fa54-403d-a383-09a4f27fccda-kube-api-access-bbh6h\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.568489 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"7476f08a-fa54-403d-a383-09a4f27fccda","Type":"ContainerDied","Data":"6965af681d16251fb13c0268cca14c6c480f5396105bd78b1a4e24b37f393fcd"} Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.568528 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.568551 4818 scope.go:117] "RemoveContainer" containerID="1d72249a0a24853a97ba44f4d6265a62ec90b1a31fcce0dfc25a3ee745ffa6ba" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.571447 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d772d6d8-be96-4256-84a5-e9f4713ad4ec","Type":"ContainerStarted","Data":"1427caf578a30c892fb575691ff95ee75649e9d4d74c48ab5a3c2380cfebf6b3"} Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.571791 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.602953 4818 scope.go:117] "RemoveContainer" containerID="5d5b24d073d0d874ef70cce146136a2725317c15b18cd5dbef5c344377a25ae6" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.608236 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.608213682 podStartE2EDuration="2.608213682s" podCreationTimestamp="2025-12-03 06:48:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:48:47.599385375 +0000 UTC m=+1285.290994137" watchObservedRunningTime="2025-12-03 06:48:47.608213682 +0000 UTC m=+1285.299822434" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.629933 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.648836 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.658705 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:48:47 crc kubenswrapper[4818]: E1203 06:48:47.659135 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7476f08a-fa54-403d-a383-09a4f27fccda" containerName="nova-metadata-log" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.659155 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="7476f08a-fa54-403d-a383-09a4f27fccda" containerName="nova-metadata-log" Dec 03 06:48:47 crc kubenswrapper[4818]: E1203 06:48:47.659192 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7476f08a-fa54-403d-a383-09a4f27fccda" containerName="nova-metadata-metadata" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.659199 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="7476f08a-fa54-403d-a383-09a4f27fccda" containerName="nova-metadata-metadata" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.659373 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="7476f08a-fa54-403d-a383-09a4f27fccda" containerName="nova-metadata-metadata" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.659409 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="7476f08a-fa54-403d-a383-09a4f27fccda" containerName="nova-metadata-log" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.660554 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.662803 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.664246 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.671827 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.738471 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " pod="openstack/nova-metadata-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.738570 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbmcm\" (UniqueName: \"kubernetes.io/projected/8aa8deb3-c59d-4150-beb4-e68529020f60-kube-api-access-vbmcm\") pod \"nova-metadata-0\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " pod="openstack/nova-metadata-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.738600 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8aa8deb3-c59d-4150-beb4-e68529020f60-logs\") pod \"nova-metadata-0\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " pod="openstack/nova-metadata-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.738622 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " pod="openstack/nova-metadata-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.738672 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-config-data\") pod \"nova-metadata-0\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " pod="openstack/nova-metadata-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.840415 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbmcm\" (UniqueName: \"kubernetes.io/projected/8aa8deb3-c59d-4150-beb4-e68529020f60-kube-api-access-vbmcm\") pod \"nova-metadata-0\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " pod="openstack/nova-metadata-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.840482 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8aa8deb3-c59d-4150-beb4-e68529020f60-logs\") pod \"nova-metadata-0\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " pod="openstack/nova-metadata-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.840516 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " pod="openstack/nova-metadata-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.840588 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-config-data\") pod \"nova-metadata-0\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " pod="openstack/nova-metadata-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.840854 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " pod="openstack/nova-metadata-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.841175 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8aa8deb3-c59d-4150-beb4-e68529020f60-logs\") pod \"nova-metadata-0\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " pod="openstack/nova-metadata-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.845674 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " pod="openstack/nova-metadata-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.846372 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " pod="openstack/nova-metadata-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.846575 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-config-data\") pod \"nova-metadata-0\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " pod="openstack/nova-metadata-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.863749 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbmcm\" (UniqueName: \"kubernetes.io/projected/8aa8deb3-c59d-4150-beb4-e68529020f60-kube-api-access-vbmcm\") pod \"nova-metadata-0\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " pod="openstack/nova-metadata-0" Dec 03 06:48:47 crc kubenswrapper[4818]: I1203 06:48:47.976769 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 06:48:48 crc kubenswrapper[4818]: I1203 06:48:48.438196 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:48:48 crc kubenswrapper[4818]: I1203 06:48:48.583056 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8aa8deb3-c59d-4150-beb4-e68529020f60","Type":"ContainerStarted","Data":"ab8337e4b0156c9798308173a72daf3ad0da1540018a259c392605601a958682"} Dec 03 06:48:48 crc kubenswrapper[4818]: I1203 06:48:48.750107 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7476f08a-fa54-403d-a383-09a4f27fccda" path="/var/lib/kubelet/pods/7476f08a-fa54-403d-a383-09a4f27fccda/volumes" Dec 03 06:48:49 crc kubenswrapper[4818]: E1203 06:48:49.047897 4818 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="231700dfd026f94d1588e43d93e8a8a7dce63750aae9af62ec8862c58010836b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 06:48:49 crc kubenswrapper[4818]: E1203 06:48:49.049493 4818 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="231700dfd026f94d1588e43d93e8a8a7dce63750aae9af62ec8862c58010836b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 06:48:49 crc kubenswrapper[4818]: E1203 06:48:49.051231 4818 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="231700dfd026f94d1588e43d93e8a8a7dce63750aae9af62ec8862c58010836b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 06:48:49 crc kubenswrapper[4818]: E1203 06:48:49.051273 4818 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="3678d8c7-48e3-4c65-9c05-48a5b96d0fed" containerName="nova-scheduler-scheduler" Dec 03 06:48:49 crc kubenswrapper[4818]: I1203 06:48:49.593438 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8aa8deb3-c59d-4150-beb4-e68529020f60","Type":"ContainerStarted","Data":"beab1d8ca08727518b06bd399677f6f228fac5d0e7db7c6f66ec3fedc32caf63"} Dec 03 06:48:49 crc kubenswrapper[4818]: I1203 06:48:49.593483 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8aa8deb3-c59d-4150-beb4-e68529020f60","Type":"ContainerStarted","Data":"af1f3345b5796f8d0b20c74bc648f214297ce4edb68bdd11ad027e991ae2237a"} Dec 03 06:48:49 crc kubenswrapper[4818]: I1203 06:48:49.613220 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.613205519 podStartE2EDuration="2.613205519s" podCreationTimestamp="2025-12-03 06:48:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:48:49.609022506 +0000 UTC m=+1287.300631258" watchObservedRunningTime="2025-12-03 06:48:49.613205519 +0000 UTC m=+1287.304814271" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.603123 4818 generic.go:334] "Generic (PLEG): container finished" podID="3678d8c7-48e3-4c65-9c05-48a5b96d0fed" containerID="231700dfd026f94d1588e43d93e8a8a7dce63750aae9af62ec8862c58010836b" exitCode=0 Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.603671 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3678d8c7-48e3-4c65-9c05-48a5b96d0fed","Type":"ContainerDied","Data":"231700dfd026f94d1588e43d93e8a8a7dce63750aae9af62ec8862c58010836b"} Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.603699 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3678d8c7-48e3-4c65-9c05-48a5b96d0fed","Type":"ContainerDied","Data":"bdc4b5eedd52166388e241cac0f2c6d773c6d55e49825a58a2fa4111c5f5a89f"} Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.603709 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdc4b5eedd52166388e241cac0f2c6d773c6d55e49825a58a2fa4111c5f5a89f" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.605344 4818 generic.go:334] "Generic (PLEG): container finished" podID="713fc838-5ff8-4799-876e-e21d870abbc0" containerID="1dac9236c1acb12ed2e9b466a699c14085a4a6115d3b106889e255449cd77aaf" exitCode=0 Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.605429 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"713fc838-5ff8-4799-876e-e21d870abbc0","Type":"ContainerDied","Data":"1dac9236c1acb12ed2e9b466a699c14085a4a6115d3b106889e255449cd77aaf"} Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.605501 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"713fc838-5ff8-4799-876e-e21d870abbc0","Type":"ContainerDied","Data":"c853299d0559d2b77e6edf8151e3caf664494d358b3eafa370220b5893511a01"} Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.605516 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c853299d0559d2b77e6edf8151e3caf664494d358b3eafa370220b5893511a01" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.650688 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.659406 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.709558 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/713fc838-5ff8-4799-876e-e21d870abbc0-combined-ca-bundle\") pod \"713fc838-5ff8-4799-876e-e21d870abbc0\" (UID: \"713fc838-5ff8-4799-876e-e21d870abbc0\") " Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.709609 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpxk2\" (UniqueName: \"kubernetes.io/projected/713fc838-5ff8-4799-876e-e21d870abbc0-kube-api-access-lpxk2\") pod \"713fc838-5ff8-4799-876e-e21d870abbc0\" (UID: \"713fc838-5ff8-4799-876e-e21d870abbc0\") " Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.709706 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/713fc838-5ff8-4799-876e-e21d870abbc0-logs\") pod \"713fc838-5ff8-4799-876e-e21d870abbc0\" (UID: \"713fc838-5ff8-4799-876e-e21d870abbc0\") " Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.709735 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/713fc838-5ff8-4799-876e-e21d870abbc0-config-data\") pod \"713fc838-5ff8-4799-876e-e21d870abbc0\" (UID: \"713fc838-5ff8-4799-876e-e21d870abbc0\") " Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.709869 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nf7f\" (UniqueName: \"kubernetes.io/projected/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-kube-api-access-5nf7f\") pod \"3678d8c7-48e3-4c65-9c05-48a5b96d0fed\" (UID: \"3678d8c7-48e3-4c65-9c05-48a5b96d0fed\") " Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.709900 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-combined-ca-bundle\") pod \"3678d8c7-48e3-4c65-9c05-48a5b96d0fed\" (UID: \"3678d8c7-48e3-4c65-9c05-48a5b96d0fed\") " Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.710033 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-config-data\") pod \"3678d8c7-48e3-4c65-9c05-48a5b96d0fed\" (UID: \"3678d8c7-48e3-4c65-9c05-48a5b96d0fed\") " Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.710213 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/713fc838-5ff8-4799-876e-e21d870abbc0-logs" (OuterVolumeSpecName: "logs") pod "713fc838-5ff8-4799-876e-e21d870abbc0" (UID: "713fc838-5ff8-4799-876e-e21d870abbc0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.710649 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/713fc838-5ff8-4799-876e-e21d870abbc0-logs\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.716260 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/713fc838-5ff8-4799-876e-e21d870abbc0-kube-api-access-lpxk2" (OuterVolumeSpecName: "kube-api-access-lpxk2") pod "713fc838-5ff8-4799-876e-e21d870abbc0" (UID: "713fc838-5ff8-4799-876e-e21d870abbc0"). InnerVolumeSpecName "kube-api-access-lpxk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.724988 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-kube-api-access-5nf7f" (OuterVolumeSpecName: "kube-api-access-5nf7f") pod "3678d8c7-48e3-4c65-9c05-48a5b96d0fed" (UID: "3678d8c7-48e3-4c65-9c05-48a5b96d0fed"). InnerVolumeSpecName "kube-api-access-5nf7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.744292 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3678d8c7-48e3-4c65-9c05-48a5b96d0fed" (UID: "3678d8c7-48e3-4c65-9c05-48a5b96d0fed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.752170 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-config-data" (OuterVolumeSpecName: "config-data") pod "3678d8c7-48e3-4c65-9c05-48a5b96d0fed" (UID: "3678d8c7-48e3-4c65-9c05-48a5b96d0fed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.773574 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/713fc838-5ff8-4799-876e-e21d870abbc0-config-data" (OuterVolumeSpecName: "config-data") pod "713fc838-5ff8-4799-876e-e21d870abbc0" (UID: "713fc838-5ff8-4799-876e-e21d870abbc0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.773651 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/713fc838-5ff8-4799-876e-e21d870abbc0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "713fc838-5ff8-4799-876e-e21d870abbc0" (UID: "713fc838-5ff8-4799-876e-e21d870abbc0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.812939 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.812983 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/713fc838-5ff8-4799-876e-e21d870abbc0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.812999 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpxk2\" (UniqueName: \"kubernetes.io/projected/713fc838-5ff8-4799-876e-e21d870abbc0-kube-api-access-lpxk2\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.813010 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/713fc838-5ff8-4799-876e-e21d870abbc0-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.813021 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nf7f\" (UniqueName: \"kubernetes.io/projected/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-kube-api-access-5nf7f\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:50 crc kubenswrapper[4818]: I1203 06:48:50.813033 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3678d8c7-48e3-4c65-9c05-48a5b96d0fed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.529663 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.614100 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.614137 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.644484 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.654887 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.716880 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.749214 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 06:48:51 crc kubenswrapper[4818]: E1203 06:48:51.749557 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="713fc838-5ff8-4799-876e-e21d870abbc0" containerName="nova-api-log" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.749573 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="713fc838-5ff8-4799-876e-e21d870abbc0" containerName="nova-api-log" Dec 03 06:48:51 crc kubenswrapper[4818]: E1203 06:48:51.749587 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="713fc838-5ff8-4799-876e-e21d870abbc0" containerName="nova-api-api" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.749593 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="713fc838-5ff8-4799-876e-e21d870abbc0" containerName="nova-api-api" Dec 03 06:48:51 crc kubenswrapper[4818]: E1203 06:48:51.749606 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3678d8c7-48e3-4c65-9c05-48a5b96d0fed" containerName="nova-scheduler-scheduler" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.749612 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="3678d8c7-48e3-4c65-9c05-48a5b96d0fed" containerName="nova-scheduler-scheduler" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.749782 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="3678d8c7-48e3-4c65-9c05-48a5b96d0fed" containerName="nova-scheduler-scheduler" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.749805 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="713fc838-5ff8-4799-876e-e21d870abbc0" containerName="nova-api-log" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.749840 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="713fc838-5ff8-4799-876e-e21d870abbc0" containerName="nova-api-api" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.750400 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.753433 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.760713 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.773791 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.785841 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.787616 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.794237 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.795581 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.830535 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d412fb3-cb4c-406e-a88c-c7b4787443fd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\") " pod="openstack/nova-api-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.830623 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d412fb3-cb4c-406e-a88c-c7b4787443fd-config-data\") pod \"nova-api-0\" (UID: \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\") " pod="openstack/nova-api-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.830729 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzw6h\" (UniqueName: \"kubernetes.io/projected/b87c588e-6d4c-49b9-87ac-1d99694ffafd-kube-api-access-zzw6h\") pod \"nova-scheduler-0\" (UID: \"b87c588e-6d4c-49b9-87ac-1d99694ffafd\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.830802 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b87c588e-6d4c-49b9-87ac-1d99694ffafd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b87c588e-6d4c-49b9-87ac-1d99694ffafd\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.830841 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b87c588e-6d4c-49b9-87ac-1d99694ffafd-config-data\") pod \"nova-scheduler-0\" (UID: \"b87c588e-6d4c-49b9-87ac-1d99694ffafd\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.830894 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d412fb3-cb4c-406e-a88c-c7b4787443fd-logs\") pod \"nova-api-0\" (UID: \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\") " pod="openstack/nova-api-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.830938 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz6zn\" (UniqueName: \"kubernetes.io/projected/7d412fb3-cb4c-406e-a88c-c7b4787443fd-kube-api-access-fz6zn\") pod \"nova-api-0\" (UID: \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\") " pod="openstack/nova-api-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.932422 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d412fb3-cb4c-406e-a88c-c7b4787443fd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\") " pod="openstack/nova-api-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.932477 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d412fb3-cb4c-406e-a88c-c7b4787443fd-config-data\") pod \"nova-api-0\" (UID: \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\") " pod="openstack/nova-api-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.932526 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzw6h\" (UniqueName: \"kubernetes.io/projected/b87c588e-6d4c-49b9-87ac-1d99694ffafd-kube-api-access-zzw6h\") pod \"nova-scheduler-0\" (UID: \"b87c588e-6d4c-49b9-87ac-1d99694ffafd\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.932566 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b87c588e-6d4c-49b9-87ac-1d99694ffafd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b87c588e-6d4c-49b9-87ac-1d99694ffafd\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.932584 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b87c588e-6d4c-49b9-87ac-1d99694ffafd-config-data\") pod \"nova-scheduler-0\" (UID: \"b87c588e-6d4c-49b9-87ac-1d99694ffafd\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.932612 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d412fb3-cb4c-406e-a88c-c7b4787443fd-logs\") pod \"nova-api-0\" (UID: \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\") " pod="openstack/nova-api-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.932647 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz6zn\" (UniqueName: \"kubernetes.io/projected/7d412fb3-cb4c-406e-a88c-c7b4787443fd-kube-api-access-fz6zn\") pod \"nova-api-0\" (UID: \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\") " pod="openstack/nova-api-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.934717 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d412fb3-cb4c-406e-a88c-c7b4787443fd-logs\") pod \"nova-api-0\" (UID: \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\") " pod="openstack/nova-api-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.940982 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b87c588e-6d4c-49b9-87ac-1d99694ffafd-config-data\") pod \"nova-scheduler-0\" (UID: \"b87c588e-6d4c-49b9-87ac-1d99694ffafd\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.942307 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d412fb3-cb4c-406e-a88c-c7b4787443fd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\") " pod="openstack/nova-api-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.942431 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b87c588e-6d4c-49b9-87ac-1d99694ffafd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b87c588e-6d4c-49b9-87ac-1d99694ffafd\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.954953 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz6zn\" (UniqueName: \"kubernetes.io/projected/7d412fb3-cb4c-406e-a88c-c7b4787443fd-kube-api-access-fz6zn\") pod \"nova-api-0\" (UID: \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\") " pod="openstack/nova-api-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.960459 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d412fb3-cb4c-406e-a88c-c7b4787443fd-config-data\") pod \"nova-api-0\" (UID: \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\") " pod="openstack/nova-api-0" Dec 03 06:48:51 crc kubenswrapper[4818]: I1203 06:48:51.964599 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzw6h\" (UniqueName: \"kubernetes.io/projected/b87c588e-6d4c-49b9-87ac-1d99694ffafd-kube-api-access-zzw6h\") pod \"nova-scheduler-0\" (UID: \"b87c588e-6d4c-49b9-87ac-1d99694ffafd\") " pod="openstack/nova-scheduler-0" Dec 03 06:48:52 crc kubenswrapper[4818]: I1203 06:48:52.163500 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 06:48:52 crc kubenswrapper[4818]: I1203 06:48:52.164425 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 06:48:52 crc kubenswrapper[4818]: I1203 06:48:52.700199 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 06:48:52 crc kubenswrapper[4818]: W1203 06:48:52.754911 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb87c588e_6d4c_49b9_87ac_1d99694ffafd.slice/crio-6c004ef4e0ca44eb5aaa64697eb9a0a71c3ffced770a317931701a236a2dd5f3 WatchSource:0}: Error finding container 6c004ef4e0ca44eb5aaa64697eb9a0a71c3ffced770a317931701a236a2dd5f3: Status 404 returned error can't find the container with id 6c004ef4e0ca44eb5aaa64697eb9a0a71c3ffced770a317931701a236a2dd5f3 Dec 03 06:48:52 crc kubenswrapper[4818]: I1203 06:48:52.759754 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3678d8c7-48e3-4c65-9c05-48a5b96d0fed" path="/var/lib/kubelet/pods/3678d8c7-48e3-4c65-9c05-48a5b96d0fed/volumes" Dec 03 06:48:52 crc kubenswrapper[4818]: I1203 06:48:52.762091 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="713fc838-5ff8-4799-876e-e21d870abbc0" path="/var/lib/kubelet/pods/713fc838-5ff8-4799-876e-e21d870abbc0/volumes" Dec 03 06:48:52 crc kubenswrapper[4818]: I1203 06:48:52.762875 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 06:48:52 crc kubenswrapper[4818]: I1203 06:48:52.977389 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 06:48:52 crc kubenswrapper[4818]: I1203 06:48:52.977455 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 06:48:53 crc kubenswrapper[4818]: I1203 06:48:53.634885 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d412fb3-cb4c-406e-a88c-c7b4787443fd","Type":"ContainerStarted","Data":"b8cd7ec7d56d8f999b3b911e243bd10aaffdadae0514351f52d72a16606ddabd"} Dec 03 06:48:53 crc kubenswrapper[4818]: I1203 06:48:53.635179 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d412fb3-cb4c-406e-a88c-c7b4787443fd","Type":"ContainerStarted","Data":"af04fa69b1fb083430371c337c587687d7272cdd989650451cba86405d5884b6"} Dec 03 06:48:53 crc kubenswrapper[4818]: I1203 06:48:53.635192 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d412fb3-cb4c-406e-a88c-c7b4787443fd","Type":"ContainerStarted","Data":"0d5ed0331e10c84b13b4ae807748d391201ad32c09afbea90a149e3b104ebd23"} Dec 03 06:48:53 crc kubenswrapper[4818]: I1203 06:48:53.638349 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b87c588e-6d4c-49b9-87ac-1d99694ffafd","Type":"ContainerStarted","Data":"8511681a19b1f575d11ea59da1ebcd4ad4a52d0a90258354449e67f91d1e8f7d"} Dec 03 06:48:53 crc kubenswrapper[4818]: I1203 06:48:53.638373 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b87c588e-6d4c-49b9-87ac-1d99694ffafd","Type":"ContainerStarted","Data":"6c004ef4e0ca44eb5aaa64697eb9a0a71c3ffced770a317931701a236a2dd5f3"} Dec 03 06:48:53 crc kubenswrapper[4818]: I1203 06:48:53.657286 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.657267719 podStartE2EDuration="2.657267719s" podCreationTimestamp="2025-12-03 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:48:53.650036662 +0000 UTC m=+1291.341645414" watchObservedRunningTime="2025-12-03 06:48:53.657267719 +0000 UTC m=+1291.348876471" Dec 03 06:48:53 crc kubenswrapper[4818]: I1203 06:48:53.675032 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.675015415 podStartE2EDuration="2.675015415s" podCreationTimestamp="2025-12-03 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:48:53.669339155 +0000 UTC m=+1291.360947907" watchObservedRunningTime="2025-12-03 06:48:53.675015415 +0000 UTC m=+1291.366624167" Dec 03 06:48:55 crc kubenswrapper[4818]: I1203 06:48:55.653913 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 06:48:55 crc kubenswrapper[4818]: I1203 06:48:55.654412 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="644cd011-5c7b-4050-85e1-2194f5cd8d93" containerName="kube-state-metrics" containerID="cri-o://f2fad1afc9124948c8ad10ceba58d3df7aaf314a58b99db8293a8339ae7a124a" gracePeriod=30 Dec 03 06:48:55 crc kubenswrapper[4818]: I1203 06:48:55.983283 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.113992 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.176736 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pl8tq\" (UniqueName: \"kubernetes.io/projected/644cd011-5c7b-4050-85e1-2194f5cd8d93-kube-api-access-pl8tq\") pod \"644cd011-5c7b-4050-85e1-2194f5cd8d93\" (UID: \"644cd011-5c7b-4050-85e1-2194f5cd8d93\") " Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.182893 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/644cd011-5c7b-4050-85e1-2194f5cd8d93-kube-api-access-pl8tq" (OuterVolumeSpecName: "kube-api-access-pl8tq") pod "644cd011-5c7b-4050-85e1-2194f5cd8d93" (UID: "644cd011-5c7b-4050-85e1-2194f5cd8d93"). InnerVolumeSpecName "kube-api-access-pl8tq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.279697 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pl8tq\" (UniqueName: \"kubernetes.io/projected/644cd011-5c7b-4050-85e1-2194f5cd8d93-kube-api-access-pl8tq\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.665877 4818 generic.go:334] "Generic (PLEG): container finished" podID="644cd011-5c7b-4050-85e1-2194f5cd8d93" containerID="f2fad1afc9124948c8ad10ceba58d3df7aaf314a58b99db8293a8339ae7a124a" exitCode=2 Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.667328 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.667336 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"644cd011-5c7b-4050-85e1-2194f5cd8d93","Type":"ContainerDied","Data":"f2fad1afc9124948c8ad10ceba58d3df7aaf314a58b99db8293a8339ae7a124a"} Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.667662 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"644cd011-5c7b-4050-85e1-2194f5cd8d93","Type":"ContainerDied","Data":"9c756d3206a98bcc761e1c0fa5a06c152c6c3fae18fcf3cc02c30fd7e7df8721"} Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.667743 4818 scope.go:117] "RemoveContainer" containerID="f2fad1afc9124948c8ad10ceba58d3df7aaf314a58b99db8293a8339ae7a124a" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.690438 4818 scope.go:117] "RemoveContainer" containerID="f2fad1afc9124948c8ad10ceba58d3df7aaf314a58b99db8293a8339ae7a124a" Dec 03 06:48:56 crc kubenswrapper[4818]: E1203 06:48:56.691192 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2fad1afc9124948c8ad10ceba58d3df7aaf314a58b99db8293a8339ae7a124a\": container with ID starting with f2fad1afc9124948c8ad10ceba58d3df7aaf314a58b99db8293a8339ae7a124a not found: ID does not exist" containerID="f2fad1afc9124948c8ad10ceba58d3df7aaf314a58b99db8293a8339ae7a124a" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.691622 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2fad1afc9124948c8ad10ceba58d3df7aaf314a58b99db8293a8339ae7a124a"} err="failed to get container status \"f2fad1afc9124948c8ad10ceba58d3df7aaf314a58b99db8293a8339ae7a124a\": rpc error: code = NotFound desc = could not find container \"f2fad1afc9124948c8ad10ceba58d3df7aaf314a58b99db8293a8339ae7a124a\": container with ID starting with f2fad1afc9124948c8ad10ceba58d3df7aaf314a58b99db8293a8339ae7a124a not found: ID does not exist" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.706802 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.723762 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.733389 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 06:48:56 crc kubenswrapper[4818]: E1203 06:48:56.733912 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="644cd011-5c7b-4050-85e1-2194f5cd8d93" containerName="kube-state-metrics" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.733932 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="644cd011-5c7b-4050-85e1-2194f5cd8d93" containerName="kube-state-metrics" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.734130 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="644cd011-5c7b-4050-85e1-2194f5cd8d93" containerName="kube-state-metrics" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.734888 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.741800 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.769693 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.770084 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.784105 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="644cd011-5c7b-4050-85e1-2194f5cd8d93" path="/var/lib/kubelet/pods/644cd011-5c7b-4050-85e1-2194f5cd8d93/volumes" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.893904 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/0d105427-cef6-4ec6-8997-08f8a94e5b56-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"0d105427-cef6-4ec6-8997-08f8a94e5b56\") " pod="openstack/kube-state-metrics-0" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.894609 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d105427-cef6-4ec6-8997-08f8a94e5b56-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"0d105427-cef6-4ec6-8997-08f8a94e5b56\") " pod="openstack/kube-state-metrics-0" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.894662 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5dtv\" (UniqueName: \"kubernetes.io/projected/0d105427-cef6-4ec6-8997-08f8a94e5b56-kube-api-access-m5dtv\") pod \"kube-state-metrics-0\" (UID: \"0d105427-cef6-4ec6-8997-08f8a94e5b56\") " pod="openstack/kube-state-metrics-0" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.894703 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d105427-cef6-4ec6-8997-08f8a94e5b56-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"0d105427-cef6-4ec6-8997-08f8a94e5b56\") " pod="openstack/kube-state-metrics-0" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.996646 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/0d105427-cef6-4ec6-8997-08f8a94e5b56-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"0d105427-cef6-4ec6-8997-08f8a94e5b56\") " pod="openstack/kube-state-metrics-0" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.996761 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d105427-cef6-4ec6-8997-08f8a94e5b56-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"0d105427-cef6-4ec6-8997-08f8a94e5b56\") " pod="openstack/kube-state-metrics-0" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.996807 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5dtv\" (UniqueName: \"kubernetes.io/projected/0d105427-cef6-4ec6-8997-08f8a94e5b56-kube-api-access-m5dtv\") pod \"kube-state-metrics-0\" (UID: \"0d105427-cef6-4ec6-8997-08f8a94e5b56\") " pod="openstack/kube-state-metrics-0" Dec 03 06:48:56 crc kubenswrapper[4818]: I1203 06:48:56.996877 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d105427-cef6-4ec6-8997-08f8a94e5b56-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"0d105427-cef6-4ec6-8997-08f8a94e5b56\") " pod="openstack/kube-state-metrics-0" Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.002334 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/0d105427-cef6-4ec6-8997-08f8a94e5b56-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"0d105427-cef6-4ec6-8997-08f8a94e5b56\") " pod="openstack/kube-state-metrics-0" Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.002628 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d105427-cef6-4ec6-8997-08f8a94e5b56-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"0d105427-cef6-4ec6-8997-08f8a94e5b56\") " pod="openstack/kube-state-metrics-0" Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.003370 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d105427-cef6-4ec6-8997-08f8a94e5b56-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"0d105427-cef6-4ec6-8997-08f8a94e5b56\") " pod="openstack/kube-state-metrics-0" Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.030950 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5dtv\" (UniqueName: \"kubernetes.io/projected/0d105427-cef6-4ec6-8997-08f8a94e5b56-kube-api-access-m5dtv\") pod \"kube-state-metrics-0\" (UID: \"0d105427-cef6-4ec6-8997-08f8a94e5b56\") " pod="openstack/kube-state-metrics-0" Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.092304 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.164690 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.535339 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.535950 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b7e28850-7e84-401f-8015-9ed869354e9b" containerName="ceilometer-central-agent" containerID="cri-o://98611168c65599f4c5e49c807527d3ba3df3c70f32abcb9b7302085e6abcaa2e" gracePeriod=30 Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.536206 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b7e28850-7e84-401f-8015-9ed869354e9b" containerName="proxy-httpd" containerID="cri-o://5f45bb410933aa6bc90857efe0045b36269cffab652eaec93fefdc045f2fb23b" gracePeriod=30 Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.536222 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b7e28850-7e84-401f-8015-9ed869354e9b" containerName="ceilometer-notification-agent" containerID="cri-o://dafe75ff8f40e67a8d5af1ac2a59f8acbfd88e249efd0dd2df3f942379dd4674" gracePeriod=30 Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.536256 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b7e28850-7e84-401f-8015-9ed869354e9b" containerName="sg-core" containerID="cri-o://b91146b25fb33b64f4db2d2d737328df7fbf72a90e372c764906b87477f71b72" gracePeriod=30 Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.571735 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.678974 4818 generic.go:334] "Generic (PLEG): container finished" podID="b7e28850-7e84-401f-8015-9ed869354e9b" containerID="5f45bb410933aa6bc90857efe0045b36269cffab652eaec93fefdc045f2fb23b" exitCode=0 Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.679033 4818 generic.go:334] "Generic (PLEG): container finished" podID="b7e28850-7e84-401f-8015-9ed869354e9b" containerID="b91146b25fb33b64f4db2d2d737328df7fbf72a90e372c764906b87477f71b72" exitCode=2 Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.679050 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7e28850-7e84-401f-8015-9ed869354e9b","Type":"ContainerDied","Data":"5f45bb410933aa6bc90857efe0045b36269cffab652eaec93fefdc045f2fb23b"} Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.679104 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7e28850-7e84-401f-8015-9ed869354e9b","Type":"ContainerDied","Data":"b91146b25fb33b64f4db2d2d737328df7fbf72a90e372c764906b87477f71b72"} Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.681272 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0d105427-cef6-4ec6-8997-08f8a94e5b56","Type":"ContainerStarted","Data":"423dbf13032ccf7d03aa26f81078e89eb86ce869aa066aac340cd2ef85f1a781"} Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.977422 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 06:48:57 crc kubenswrapper[4818]: I1203 06:48:57.977491 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 06:48:58 crc kubenswrapper[4818]: I1203 06:48:58.693554 4818 generic.go:334] "Generic (PLEG): container finished" podID="b7e28850-7e84-401f-8015-9ed869354e9b" containerID="98611168c65599f4c5e49c807527d3ba3df3c70f32abcb9b7302085e6abcaa2e" exitCode=0 Dec 03 06:48:58 crc kubenswrapper[4818]: I1203 06:48:58.693601 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7e28850-7e84-401f-8015-9ed869354e9b","Type":"ContainerDied","Data":"98611168c65599f4c5e49c807527d3ba3df3c70f32abcb9b7302085e6abcaa2e"} Dec 03 06:48:58 crc kubenswrapper[4818]: I1203 06:48:58.695375 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0d105427-cef6-4ec6-8997-08f8a94e5b56","Type":"ContainerStarted","Data":"b57d1cee0294a49943651b6ed75a2b046a63d34e1735952088329f22527fb109"} Dec 03 06:48:58 crc kubenswrapper[4818]: I1203 06:48:58.695546 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 06:48:58 crc kubenswrapper[4818]: I1203 06:48:58.724650 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.368144523 podStartE2EDuration="2.724623542s" podCreationTimestamp="2025-12-03 06:48:56 +0000 UTC" firstStartedPulling="2025-12-03 06:48:57.579659883 +0000 UTC m=+1295.271268635" lastFinishedPulling="2025-12-03 06:48:57.936138902 +0000 UTC m=+1295.627747654" observedRunningTime="2025-12-03 06:48:58.714675538 +0000 UTC m=+1296.406284330" watchObservedRunningTime="2025-12-03 06:48:58.724623542 +0000 UTC m=+1296.416232334" Dec 03 06:48:58 crc kubenswrapper[4818]: I1203 06:48:58.989163 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8aa8deb3-c59d-4150-beb4-e68529020f60" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 06:48:58 crc kubenswrapper[4818]: I1203 06:48:58.989512 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8aa8deb3-c59d-4150-beb4-e68529020f60" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.533401 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.537911 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46bh7\" (UniqueName: \"kubernetes.io/projected/b7e28850-7e84-401f-8015-9ed869354e9b-kube-api-access-46bh7\") pod \"b7e28850-7e84-401f-8015-9ed869354e9b\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.537975 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-sg-core-conf-yaml\") pod \"b7e28850-7e84-401f-8015-9ed869354e9b\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.538001 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-scripts\") pod \"b7e28850-7e84-401f-8015-9ed869354e9b\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.538033 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-config-data\") pod \"b7e28850-7e84-401f-8015-9ed869354e9b\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.538063 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7e28850-7e84-401f-8015-9ed869354e9b-run-httpd\") pod \"b7e28850-7e84-401f-8015-9ed869354e9b\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.538145 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7e28850-7e84-401f-8015-9ed869354e9b-log-httpd\") pod \"b7e28850-7e84-401f-8015-9ed869354e9b\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.538175 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-combined-ca-bundle\") pod \"b7e28850-7e84-401f-8015-9ed869354e9b\" (UID: \"b7e28850-7e84-401f-8015-9ed869354e9b\") " Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.538366 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7e28850-7e84-401f-8015-9ed869354e9b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b7e28850-7e84-401f-8015-9ed869354e9b" (UID: "b7e28850-7e84-401f-8015-9ed869354e9b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.538521 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7e28850-7e84-401f-8015-9ed869354e9b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b7e28850-7e84-401f-8015-9ed869354e9b" (UID: "b7e28850-7e84-401f-8015-9ed869354e9b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.538607 4818 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7e28850-7e84-401f-8015-9ed869354e9b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.545215 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7e28850-7e84-401f-8015-9ed869354e9b-kube-api-access-46bh7" (OuterVolumeSpecName: "kube-api-access-46bh7") pod "b7e28850-7e84-401f-8015-9ed869354e9b" (UID: "b7e28850-7e84-401f-8015-9ed869354e9b"). InnerVolumeSpecName "kube-api-access-46bh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.550716 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-scripts" (OuterVolumeSpecName: "scripts") pod "b7e28850-7e84-401f-8015-9ed869354e9b" (UID: "b7e28850-7e84-401f-8015-9ed869354e9b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.579892 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b7e28850-7e84-401f-8015-9ed869354e9b" (UID: "b7e28850-7e84-401f-8015-9ed869354e9b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.640013 4818 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7e28850-7e84-401f-8015-9ed869354e9b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.640050 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46bh7\" (UniqueName: \"kubernetes.io/projected/b7e28850-7e84-401f-8015-9ed869354e9b-kube-api-access-46bh7\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.640065 4818 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.640076 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.654797 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b7e28850-7e84-401f-8015-9ed869354e9b" (UID: "b7e28850-7e84-401f-8015-9ed869354e9b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.672542 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-config-data" (OuterVolumeSpecName: "config-data") pod "b7e28850-7e84-401f-8015-9ed869354e9b" (UID: "b7e28850-7e84-401f-8015-9ed869354e9b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.710379 4818 generic.go:334] "Generic (PLEG): container finished" podID="b7e28850-7e84-401f-8015-9ed869354e9b" containerID="dafe75ff8f40e67a8d5af1ac2a59f8acbfd88e249efd0dd2df3f942379dd4674" exitCode=0 Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.710454 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.710466 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7e28850-7e84-401f-8015-9ed869354e9b","Type":"ContainerDied","Data":"dafe75ff8f40e67a8d5af1ac2a59f8acbfd88e249efd0dd2df3f942379dd4674"} Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.711805 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7e28850-7e84-401f-8015-9ed869354e9b","Type":"ContainerDied","Data":"cf324af06b14813bb47fe0eb305607db0ce82a33e812c0a1d7627d14b99ab09d"} Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.711847 4818 scope.go:117] "RemoveContainer" containerID="5f45bb410933aa6bc90857efe0045b36269cffab652eaec93fefdc045f2fb23b" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.738509 4818 scope.go:117] "RemoveContainer" containerID="b91146b25fb33b64f4db2d2d737328df7fbf72a90e372c764906b87477f71b72" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.750067 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.750187 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e28850-7e84-401f-8015-9ed869354e9b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.755927 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.773305 4818 scope.go:117] "RemoveContainer" containerID="dafe75ff8f40e67a8d5af1ac2a59f8acbfd88e249efd0dd2df3f942379dd4674" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.781452 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.801513 4818 scope.go:117] "RemoveContainer" containerID="98611168c65599f4c5e49c807527d3ba3df3c70f32abcb9b7302085e6abcaa2e" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.801695 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:48:59 crc kubenswrapper[4818]: E1203 06:48:59.802119 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7e28850-7e84-401f-8015-9ed869354e9b" containerName="sg-core" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.807060 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7e28850-7e84-401f-8015-9ed869354e9b" containerName="sg-core" Dec 03 06:48:59 crc kubenswrapper[4818]: E1203 06:48:59.807157 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7e28850-7e84-401f-8015-9ed869354e9b" containerName="proxy-httpd" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.807168 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7e28850-7e84-401f-8015-9ed869354e9b" containerName="proxy-httpd" Dec 03 06:48:59 crc kubenswrapper[4818]: E1203 06:48:59.807188 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7e28850-7e84-401f-8015-9ed869354e9b" containerName="ceilometer-central-agent" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.807194 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7e28850-7e84-401f-8015-9ed869354e9b" containerName="ceilometer-central-agent" Dec 03 06:48:59 crc kubenswrapper[4818]: E1203 06:48:59.807230 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7e28850-7e84-401f-8015-9ed869354e9b" containerName="ceilometer-notification-agent" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.807236 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7e28850-7e84-401f-8015-9ed869354e9b" containerName="ceilometer-notification-agent" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.807556 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7e28850-7e84-401f-8015-9ed869354e9b" containerName="proxy-httpd" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.807582 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7e28850-7e84-401f-8015-9ed869354e9b" containerName="ceilometer-notification-agent" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.807598 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7e28850-7e84-401f-8015-9ed869354e9b" containerName="ceilometer-central-agent" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.807607 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7e28850-7e84-401f-8015-9ed869354e9b" containerName="sg-core" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.809830 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.809917 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.812741 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.812975 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.813278 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.840438 4818 scope.go:117] "RemoveContainer" containerID="5f45bb410933aa6bc90857efe0045b36269cffab652eaec93fefdc045f2fb23b" Dec 03 06:48:59 crc kubenswrapper[4818]: E1203 06:48:59.840823 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f45bb410933aa6bc90857efe0045b36269cffab652eaec93fefdc045f2fb23b\": container with ID starting with 5f45bb410933aa6bc90857efe0045b36269cffab652eaec93fefdc045f2fb23b not found: ID does not exist" containerID="5f45bb410933aa6bc90857efe0045b36269cffab652eaec93fefdc045f2fb23b" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.840852 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f45bb410933aa6bc90857efe0045b36269cffab652eaec93fefdc045f2fb23b"} err="failed to get container status \"5f45bb410933aa6bc90857efe0045b36269cffab652eaec93fefdc045f2fb23b\": rpc error: code = NotFound desc = could not find container \"5f45bb410933aa6bc90857efe0045b36269cffab652eaec93fefdc045f2fb23b\": container with ID starting with 5f45bb410933aa6bc90857efe0045b36269cffab652eaec93fefdc045f2fb23b not found: ID does not exist" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.840870 4818 scope.go:117] "RemoveContainer" containerID="b91146b25fb33b64f4db2d2d737328df7fbf72a90e372c764906b87477f71b72" Dec 03 06:48:59 crc kubenswrapper[4818]: E1203 06:48:59.841241 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b91146b25fb33b64f4db2d2d737328df7fbf72a90e372c764906b87477f71b72\": container with ID starting with b91146b25fb33b64f4db2d2d737328df7fbf72a90e372c764906b87477f71b72 not found: ID does not exist" containerID="b91146b25fb33b64f4db2d2d737328df7fbf72a90e372c764906b87477f71b72" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.841276 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b91146b25fb33b64f4db2d2d737328df7fbf72a90e372c764906b87477f71b72"} err="failed to get container status \"b91146b25fb33b64f4db2d2d737328df7fbf72a90e372c764906b87477f71b72\": rpc error: code = NotFound desc = could not find container \"b91146b25fb33b64f4db2d2d737328df7fbf72a90e372c764906b87477f71b72\": container with ID starting with b91146b25fb33b64f4db2d2d737328df7fbf72a90e372c764906b87477f71b72 not found: ID does not exist" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.841300 4818 scope.go:117] "RemoveContainer" containerID="dafe75ff8f40e67a8d5af1ac2a59f8acbfd88e249efd0dd2df3f942379dd4674" Dec 03 06:48:59 crc kubenswrapper[4818]: E1203 06:48:59.841550 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dafe75ff8f40e67a8d5af1ac2a59f8acbfd88e249efd0dd2df3f942379dd4674\": container with ID starting with dafe75ff8f40e67a8d5af1ac2a59f8acbfd88e249efd0dd2df3f942379dd4674 not found: ID does not exist" containerID="dafe75ff8f40e67a8d5af1ac2a59f8acbfd88e249efd0dd2df3f942379dd4674" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.841577 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dafe75ff8f40e67a8d5af1ac2a59f8acbfd88e249efd0dd2df3f942379dd4674"} err="failed to get container status \"dafe75ff8f40e67a8d5af1ac2a59f8acbfd88e249efd0dd2df3f942379dd4674\": rpc error: code = NotFound desc = could not find container \"dafe75ff8f40e67a8d5af1ac2a59f8acbfd88e249efd0dd2df3f942379dd4674\": container with ID starting with dafe75ff8f40e67a8d5af1ac2a59f8acbfd88e249efd0dd2df3f942379dd4674 not found: ID does not exist" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.841597 4818 scope.go:117] "RemoveContainer" containerID="98611168c65599f4c5e49c807527d3ba3df3c70f32abcb9b7302085e6abcaa2e" Dec 03 06:48:59 crc kubenswrapper[4818]: E1203 06:48:59.841774 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98611168c65599f4c5e49c807527d3ba3df3c70f32abcb9b7302085e6abcaa2e\": container with ID starting with 98611168c65599f4c5e49c807527d3ba3df3c70f32abcb9b7302085e6abcaa2e not found: ID does not exist" containerID="98611168c65599f4c5e49c807527d3ba3df3c70f32abcb9b7302085e6abcaa2e" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.841793 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98611168c65599f4c5e49c807527d3ba3df3c70f32abcb9b7302085e6abcaa2e"} err="failed to get container status \"98611168c65599f4c5e49c807527d3ba3df3c70f32abcb9b7302085e6abcaa2e\": rpc error: code = NotFound desc = could not find container \"98611168c65599f4c5e49c807527d3ba3df3c70f32abcb9b7302085e6abcaa2e\": container with ID starting with 98611168c65599f4c5e49c807527d3ba3df3c70f32abcb9b7302085e6abcaa2e not found: ID does not exist" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.851611 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-scripts\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.851867 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.851995 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ccb6887-a207-45df-a993-d7e65238b9a8-run-httpd\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.852113 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.852197 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-config-data\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.852337 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ccb6887-a207-45df-a993-d7e65238b9a8-log-httpd\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.852428 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.852522 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpnbr\" (UniqueName: \"kubernetes.io/projected/4ccb6887-a207-45df-a993-d7e65238b9a8-kube-api-access-wpnbr\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.954057 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpnbr\" (UniqueName: \"kubernetes.io/projected/4ccb6887-a207-45df-a993-d7e65238b9a8-kube-api-access-wpnbr\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.954365 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-scripts\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.954432 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.954455 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ccb6887-a207-45df-a993-d7e65238b9a8-run-httpd\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.954498 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.954517 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-config-data\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.954537 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ccb6887-a207-45df-a993-d7e65238b9a8-log-httpd\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.954572 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.959107 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ccb6887-a207-45df-a993-d7e65238b9a8-log-httpd\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.959480 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.959620 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.959774 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ccb6887-a207-45df-a993-d7e65238b9a8-run-httpd\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.960766 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-config-data\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.961331 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.961595 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-scripts\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:48:59 crc kubenswrapper[4818]: I1203 06:48:59.977630 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpnbr\" (UniqueName: \"kubernetes.io/projected/4ccb6887-a207-45df-a993-d7e65238b9a8-kube-api-access-wpnbr\") pod \"ceilometer-0\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " pod="openstack/ceilometer-0" Dec 03 06:49:00 crc kubenswrapper[4818]: I1203 06:49:00.138726 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:49:00 crc kubenswrapper[4818]: I1203 06:49:00.627211 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:49:00 crc kubenswrapper[4818]: I1203 06:49:00.721885 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ccb6887-a207-45df-a993-d7e65238b9a8","Type":"ContainerStarted","Data":"784833c48527f8c5b0853ee417f8aef628a9820eadf727dca101d0833e3ade06"} Dec 03 06:49:00 crc kubenswrapper[4818]: I1203 06:49:00.752611 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7e28850-7e84-401f-8015-9ed869354e9b" path="/var/lib/kubelet/pods/b7e28850-7e84-401f-8015-9ed869354e9b/volumes" Dec 03 06:49:01 crc kubenswrapper[4818]: I1203 06:49:01.733889 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ccb6887-a207-45df-a993-d7e65238b9a8","Type":"ContainerStarted","Data":"b35ca1ef7881533da7ef2a5095163390a08d73582c41b819e855bcb5a4bf71e9"} Dec 03 06:49:02 crc kubenswrapper[4818]: I1203 06:49:02.164223 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 06:49:02 crc kubenswrapper[4818]: I1203 06:49:02.165500 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 06:49:02 crc kubenswrapper[4818]: I1203 06:49:02.165952 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 06:49:02 crc kubenswrapper[4818]: I1203 06:49:02.201652 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 06:49:02 crc kubenswrapper[4818]: I1203 06:49:02.755040 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ccb6887-a207-45df-a993-d7e65238b9a8","Type":"ContainerStarted","Data":"f56727a2c66c4ee4e1ab1f8f696efa618028690b75d1b5097835730b77425bcb"} Dec 03 06:49:02 crc kubenswrapper[4818]: I1203 06:49:02.789085 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 06:49:03 crc kubenswrapper[4818]: I1203 06:49:03.247940 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7d412fb3-cb4c-406e-a88c-c7b4787443fd" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 06:49:03 crc kubenswrapper[4818]: I1203 06:49:03.247978 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7d412fb3-cb4c-406e-a88c-c7b4787443fd" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 06:49:03 crc kubenswrapper[4818]: I1203 06:49:03.762302 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ccb6887-a207-45df-a993-d7e65238b9a8","Type":"ContainerStarted","Data":"695ef63c2a501763ef29bb40dd5bb218b57bede54065e47e786e9872b759af94"} Dec 03 06:49:04 crc kubenswrapper[4818]: I1203 06:49:04.773490 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ccb6887-a207-45df-a993-d7e65238b9a8","Type":"ContainerStarted","Data":"82d4f1d7f39e525e1e592f9eb128d604910b2e24aa3a7da1fe38631d379d11a6"} Dec 03 06:49:04 crc kubenswrapper[4818]: I1203 06:49:04.774152 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 06:49:04 crc kubenswrapper[4818]: I1203 06:49:04.801755 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.395494424 podStartE2EDuration="5.801733417s" podCreationTimestamp="2025-12-03 06:48:59 +0000 UTC" firstStartedPulling="2025-12-03 06:49:00.631564018 +0000 UTC m=+1298.323172770" lastFinishedPulling="2025-12-03 06:49:04.037803011 +0000 UTC m=+1301.729411763" observedRunningTime="2025-12-03 06:49:04.795116373 +0000 UTC m=+1302.486725145" watchObservedRunningTime="2025-12-03 06:49:04.801733417 +0000 UTC m=+1302.493342179" Dec 03 06:49:07 crc kubenswrapper[4818]: I1203 06:49:07.102592 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 06:49:07 crc kubenswrapper[4818]: I1203 06:49:07.986337 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 06:49:07 crc kubenswrapper[4818]: I1203 06:49:07.988191 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 06:49:07 crc kubenswrapper[4818]: I1203 06:49:07.991695 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 06:49:08 crc kubenswrapper[4818]: I1203 06:49:08.817241 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 06:49:09 crc kubenswrapper[4818]: I1203 06:49:09.823135 4818 generic.go:334] "Generic (PLEG): container finished" podID="a33f783e-79d3-41be-93e7-c703833c95aa" containerID="935aa46d2cd7f030179cdc34fbc5116fc6332f35e7dedac191ef26bf7889e304" exitCode=137 Dec 03 06:49:09 crc kubenswrapper[4818]: I1203 06:49:09.823224 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a33f783e-79d3-41be-93e7-c703833c95aa","Type":"ContainerDied","Data":"935aa46d2cd7f030179cdc34fbc5116fc6332f35e7dedac191ef26bf7889e304"} Dec 03 06:49:09 crc kubenswrapper[4818]: I1203 06:49:09.971830 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.102318 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33f783e-79d3-41be-93e7-c703833c95aa-config-data\") pod \"a33f783e-79d3-41be-93e7-c703833c95aa\" (UID: \"a33f783e-79d3-41be-93e7-c703833c95aa\") " Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.102429 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzh8n\" (UniqueName: \"kubernetes.io/projected/a33f783e-79d3-41be-93e7-c703833c95aa-kube-api-access-qzh8n\") pod \"a33f783e-79d3-41be-93e7-c703833c95aa\" (UID: \"a33f783e-79d3-41be-93e7-c703833c95aa\") " Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.102521 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33f783e-79d3-41be-93e7-c703833c95aa-combined-ca-bundle\") pod \"a33f783e-79d3-41be-93e7-c703833c95aa\" (UID: \"a33f783e-79d3-41be-93e7-c703833c95aa\") " Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.108241 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a33f783e-79d3-41be-93e7-c703833c95aa-kube-api-access-qzh8n" (OuterVolumeSpecName: "kube-api-access-qzh8n") pod "a33f783e-79d3-41be-93e7-c703833c95aa" (UID: "a33f783e-79d3-41be-93e7-c703833c95aa"). InnerVolumeSpecName "kube-api-access-qzh8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.129363 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a33f783e-79d3-41be-93e7-c703833c95aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a33f783e-79d3-41be-93e7-c703833c95aa" (UID: "a33f783e-79d3-41be-93e7-c703833c95aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.142076 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a33f783e-79d3-41be-93e7-c703833c95aa-config-data" (OuterVolumeSpecName: "config-data") pod "a33f783e-79d3-41be-93e7-c703833c95aa" (UID: "a33f783e-79d3-41be-93e7-c703833c95aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.204980 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33f783e-79d3-41be-93e7-c703833c95aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.205034 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33f783e-79d3-41be-93e7-c703833c95aa-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.205054 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzh8n\" (UniqueName: \"kubernetes.io/projected/a33f783e-79d3-41be-93e7-c703833c95aa-kube-api-access-qzh8n\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.833943 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a33f783e-79d3-41be-93e7-c703833c95aa","Type":"ContainerDied","Data":"19c47281cf174949c1c92e7ba8a39ee3f204f81ae5442baaa826d596dbfd3ab1"} Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.833969 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.834358 4818 scope.go:117] "RemoveContainer" containerID="935aa46d2cd7f030179cdc34fbc5116fc6332f35e7dedac191ef26bf7889e304" Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.861334 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.899962 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.914497 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 06:49:10 crc kubenswrapper[4818]: E1203 06:49:10.914959 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a33f783e-79d3-41be-93e7-c703833c95aa" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.914984 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a33f783e-79d3-41be-93e7-c703833c95aa" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.915236 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a33f783e-79d3-41be-93e7-c703833c95aa" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.915927 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.916024 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.929422 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.929585 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 03 06:49:10 crc kubenswrapper[4818]: I1203 06:49:10.929775 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.031999 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1f0f202-1c95-48ed-96fa-1606b2c81a6f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1f0f202-1c95-48ed-96fa-1606b2c81a6f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.032055 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1f0f202-1c95-48ed-96fa-1606b2c81a6f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1f0f202-1c95-48ed-96fa-1606b2c81a6f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.032157 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1f0f202-1c95-48ed-96fa-1606b2c81a6f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1f0f202-1c95-48ed-96fa-1606b2c81a6f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.032180 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgs5t\" (UniqueName: \"kubernetes.io/projected/b1f0f202-1c95-48ed-96fa-1606b2c81a6f-kube-api-access-pgs5t\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1f0f202-1c95-48ed-96fa-1606b2c81a6f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.032545 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1f0f202-1c95-48ed-96fa-1606b2c81a6f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1f0f202-1c95-48ed-96fa-1606b2c81a6f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.134367 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1f0f202-1c95-48ed-96fa-1606b2c81a6f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1f0f202-1c95-48ed-96fa-1606b2c81a6f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.134475 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1f0f202-1c95-48ed-96fa-1606b2c81a6f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1f0f202-1c95-48ed-96fa-1606b2c81a6f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.134507 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgs5t\" (UniqueName: \"kubernetes.io/projected/b1f0f202-1c95-48ed-96fa-1606b2c81a6f-kube-api-access-pgs5t\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1f0f202-1c95-48ed-96fa-1606b2c81a6f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.134616 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1f0f202-1c95-48ed-96fa-1606b2c81a6f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1f0f202-1c95-48ed-96fa-1606b2c81a6f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.134679 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1f0f202-1c95-48ed-96fa-1606b2c81a6f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1f0f202-1c95-48ed-96fa-1606b2c81a6f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.139202 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1f0f202-1c95-48ed-96fa-1606b2c81a6f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1f0f202-1c95-48ed-96fa-1606b2c81a6f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.139655 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1f0f202-1c95-48ed-96fa-1606b2c81a6f-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1f0f202-1c95-48ed-96fa-1606b2c81a6f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.140218 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1f0f202-1c95-48ed-96fa-1606b2c81a6f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1f0f202-1c95-48ed-96fa-1606b2c81a6f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.141535 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1f0f202-1c95-48ed-96fa-1606b2c81a6f-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1f0f202-1c95-48ed-96fa-1606b2c81a6f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.157624 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgs5t\" (UniqueName: \"kubernetes.io/projected/b1f0f202-1c95-48ed-96fa-1606b2c81a6f-kube-api-access-pgs5t\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1f0f202-1c95-48ed-96fa-1606b2c81a6f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.250393 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.747366 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 06:49:11 crc kubenswrapper[4818]: I1203 06:49:11.854718 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b1f0f202-1c95-48ed-96fa-1606b2c81a6f","Type":"ContainerStarted","Data":"8e90572791471bf4699c160a8c5ef8c26de52687f7c156a581df20843e196cb7"} Dec 03 06:49:12 crc kubenswrapper[4818]: I1203 06:49:12.181594 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 06:49:12 crc kubenswrapper[4818]: I1203 06:49:12.181997 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 06:49:12 crc kubenswrapper[4818]: I1203 06:49:12.184778 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 06:49:12 crc kubenswrapper[4818]: I1203 06:49:12.188156 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 06:49:12 crc kubenswrapper[4818]: I1203 06:49:12.754886 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a33f783e-79d3-41be-93e7-c703833c95aa" path="/var/lib/kubelet/pods/a33f783e-79d3-41be-93e7-c703833c95aa/volumes" Dec 03 06:49:12 crc kubenswrapper[4818]: I1203 06:49:12.866418 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b1f0f202-1c95-48ed-96fa-1606b2c81a6f","Type":"ContainerStarted","Data":"b0902e347aa380dafb4f019a689182794e50cc05c324e4da6d3224453f721432"} Dec 03 06:49:12 crc kubenswrapper[4818]: I1203 06:49:12.867307 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 06:49:12 crc kubenswrapper[4818]: I1203 06:49:12.870609 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 06:49:12 crc kubenswrapper[4818]: I1203 06:49:12.898959 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.898936627 podStartE2EDuration="2.898936627s" podCreationTimestamp="2025-12-03 06:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:49:12.887165966 +0000 UTC m=+1310.578774728" watchObservedRunningTime="2025-12-03 06:49:12.898936627 +0000 UTC m=+1310.590545389" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.111723 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bsnnk"] Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.113449 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.122748 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bsnnk"] Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.201797 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.201868 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.201925 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.201982 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g2mw\" (UniqueName: \"kubernetes.io/projected/0b06c002-0804-463d-84ca-0cc7fa8453b8-kube-api-access-6g2mw\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.202016 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-config\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.202039 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.302267 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.302360 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.303202 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.303294 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g2mw\" (UniqueName: \"kubernetes.io/projected/0b06c002-0804-463d-84ca-0cc7fa8453b8-kube-api-access-6g2mw\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.303342 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-config\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.303400 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.303561 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.303584 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.304602 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.304718 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.304753 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.305038 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.305258 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-config\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.321156 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g2mw\" (UniqueName: \"kubernetes.io/projected/0b06c002-0804-463d-84ca-0cc7fa8453b8-kube-api-access-6g2mw\") pod \"dnsmasq-dns-89c5cd4d5-bsnnk\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.435597 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:13 crc kubenswrapper[4818]: W1203 06:49:13.929607 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b06c002_0804_463d_84ca_0cc7fa8453b8.slice/crio-a447fd0071848427583a78d064adcc0467026738f3279f409b2378b320625636 WatchSource:0}: Error finding container a447fd0071848427583a78d064adcc0467026738f3279f409b2378b320625636: Status 404 returned error can't find the container with id a447fd0071848427583a78d064adcc0467026738f3279f409b2378b320625636 Dec 03 06:49:13 crc kubenswrapper[4818]: I1203 06:49:13.939872 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bsnnk"] Dec 03 06:49:14 crc kubenswrapper[4818]: I1203 06:49:14.886305 4818 generic.go:334] "Generic (PLEG): container finished" podID="0b06c002-0804-463d-84ca-0cc7fa8453b8" containerID="e484eb2f82618daab92823e31d09951aa822a64bcf65939bc3dfb37d798dc5a6" exitCode=0 Dec 03 06:49:14 crc kubenswrapper[4818]: I1203 06:49:14.887184 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" event={"ID":"0b06c002-0804-463d-84ca-0cc7fa8453b8","Type":"ContainerDied","Data":"e484eb2f82618daab92823e31d09951aa822a64bcf65939bc3dfb37d798dc5a6"} Dec 03 06:49:14 crc kubenswrapper[4818]: I1203 06:49:14.887228 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" event={"ID":"0b06c002-0804-463d-84ca-0cc7fa8453b8","Type":"ContainerStarted","Data":"a447fd0071848427583a78d064adcc0467026738f3279f409b2378b320625636"} Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.090117 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.093556 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerName="proxy-httpd" containerID="cri-o://82d4f1d7f39e525e1e592f9eb128d604910b2e24aa3a7da1fe38631d379d11a6" gracePeriod=30 Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.093701 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerName="sg-core" containerID="cri-o://695ef63c2a501763ef29bb40dd5bb218b57bede54065e47e786e9872b759af94" gracePeriod=30 Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.093840 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerName="ceilometer-notification-agent" containerID="cri-o://f56727a2c66c4ee4e1ab1f8f696efa618028690b75d1b5097835730b77425bcb" gracePeriod=30 Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.093293 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerName="ceilometer-central-agent" containerID="cri-o://b35ca1ef7881533da7ef2a5095163390a08d73582c41b819e855bcb5a4bf71e9" gracePeriod=30 Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.102588 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.195:3000/\": EOF" Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.507571 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.899364 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" event={"ID":"0b06c002-0804-463d-84ca-0cc7fa8453b8","Type":"ContainerStarted","Data":"3077c286134324a6da59bb654d265558445cd43dc0afd601b184056ae709752b"} Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.899444 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.902348 4818 generic.go:334] "Generic (PLEG): container finished" podID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerID="82d4f1d7f39e525e1e592f9eb128d604910b2e24aa3a7da1fe38631d379d11a6" exitCode=0 Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.902378 4818 generic.go:334] "Generic (PLEG): container finished" podID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerID="695ef63c2a501763ef29bb40dd5bb218b57bede54065e47e786e9872b759af94" exitCode=2 Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.902387 4818 generic.go:334] "Generic (PLEG): container finished" podID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerID="b35ca1ef7881533da7ef2a5095163390a08d73582c41b819e855bcb5a4bf71e9" exitCode=0 Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.902548 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7d412fb3-cb4c-406e-a88c-c7b4787443fd" containerName="nova-api-log" containerID="cri-o://af04fa69b1fb083430371c337c587687d7272cdd989650451cba86405d5884b6" gracePeriod=30 Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.902618 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ccb6887-a207-45df-a993-d7e65238b9a8","Type":"ContainerDied","Data":"82d4f1d7f39e525e1e592f9eb128d604910b2e24aa3a7da1fe38631d379d11a6"} Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.902639 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ccb6887-a207-45df-a993-d7e65238b9a8","Type":"ContainerDied","Data":"695ef63c2a501763ef29bb40dd5bb218b57bede54065e47e786e9872b759af94"} Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.902648 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ccb6887-a207-45df-a993-d7e65238b9a8","Type":"ContainerDied","Data":"b35ca1ef7881533da7ef2a5095163390a08d73582c41b819e855bcb5a4bf71e9"} Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.902692 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7d412fb3-cb4c-406e-a88c-c7b4787443fd" containerName="nova-api-api" containerID="cri-o://b8cd7ec7d56d8f999b3b911e243bd10aaffdadae0514351f52d72a16606ddabd" gracePeriod=30 Dec 03 06:49:15 crc kubenswrapper[4818]: I1203 06:49:15.925088 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" podStartSLOduration=2.925064549 podStartE2EDuration="2.925064549s" podCreationTimestamp="2025-12-03 06:49:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:49:15.915930863 +0000 UTC m=+1313.607539625" watchObservedRunningTime="2025-12-03 06:49:15.925064549 +0000 UTC m=+1313.616673311" Dec 03 06:49:16 crc kubenswrapper[4818]: I1203 06:49:16.251192 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:16 crc kubenswrapper[4818]: I1203 06:49:16.913697 4818 generic.go:334] "Generic (PLEG): container finished" podID="7d412fb3-cb4c-406e-a88c-c7b4787443fd" containerID="af04fa69b1fb083430371c337c587687d7272cdd989650451cba86405d5884b6" exitCode=143 Dec 03 06:49:16 crc kubenswrapper[4818]: I1203 06:49:16.913780 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d412fb3-cb4c-406e-a88c-c7b4787443fd","Type":"ContainerDied","Data":"af04fa69b1fb083430371c337c587687d7272cdd989650451cba86405d5884b6"} Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.442144 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.496835 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-combined-ca-bundle\") pod \"4ccb6887-a207-45df-a993-d7e65238b9a8\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.496887 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpnbr\" (UniqueName: \"kubernetes.io/projected/4ccb6887-a207-45df-a993-d7e65238b9a8-kube-api-access-wpnbr\") pod \"4ccb6887-a207-45df-a993-d7e65238b9a8\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.496968 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-sg-core-conf-yaml\") pod \"4ccb6887-a207-45df-a993-d7e65238b9a8\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.496994 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-ceilometer-tls-certs\") pod \"4ccb6887-a207-45df-a993-d7e65238b9a8\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.497035 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ccb6887-a207-45df-a993-d7e65238b9a8-log-httpd\") pod \"4ccb6887-a207-45df-a993-d7e65238b9a8\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.497106 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-scripts\") pod \"4ccb6887-a207-45df-a993-d7e65238b9a8\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.497128 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-config-data\") pod \"4ccb6887-a207-45df-a993-d7e65238b9a8\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.497161 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ccb6887-a207-45df-a993-d7e65238b9a8-run-httpd\") pod \"4ccb6887-a207-45df-a993-d7e65238b9a8\" (UID: \"4ccb6887-a207-45df-a993-d7e65238b9a8\") " Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.498917 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ccb6887-a207-45df-a993-d7e65238b9a8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4ccb6887-a207-45df-a993-d7e65238b9a8" (UID: "4ccb6887-a207-45df-a993-d7e65238b9a8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.498932 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ccb6887-a207-45df-a993-d7e65238b9a8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4ccb6887-a207-45df-a993-d7e65238b9a8" (UID: "4ccb6887-a207-45df-a993-d7e65238b9a8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.499452 4818 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ccb6887-a207-45df-a993-d7e65238b9a8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.499471 4818 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ccb6887-a207-45df-a993-d7e65238b9a8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.504319 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-scripts" (OuterVolumeSpecName: "scripts") pod "4ccb6887-a207-45df-a993-d7e65238b9a8" (UID: "4ccb6887-a207-45df-a993-d7e65238b9a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.504401 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ccb6887-a207-45df-a993-d7e65238b9a8-kube-api-access-wpnbr" (OuterVolumeSpecName: "kube-api-access-wpnbr") pod "4ccb6887-a207-45df-a993-d7e65238b9a8" (UID: "4ccb6887-a207-45df-a993-d7e65238b9a8"). InnerVolumeSpecName "kube-api-access-wpnbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.530915 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4ccb6887-a207-45df-a993-d7e65238b9a8" (UID: "4ccb6887-a207-45df-a993-d7e65238b9a8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.553568 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "4ccb6887-a207-45df-a993-d7e65238b9a8" (UID: "4ccb6887-a207-45df-a993-d7e65238b9a8"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.588360 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ccb6887-a207-45df-a993-d7e65238b9a8" (UID: "4ccb6887-a207-45df-a993-d7e65238b9a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.601966 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.602002 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.602018 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpnbr\" (UniqueName: \"kubernetes.io/projected/4ccb6887-a207-45df-a993-d7e65238b9a8-kube-api-access-wpnbr\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.602030 4818 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.602041 4818 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.608135 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-config-data" (OuterVolumeSpecName: "config-data") pod "4ccb6887-a207-45df-a993-d7e65238b9a8" (UID: "4ccb6887-a207-45df-a993-d7e65238b9a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.703469 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ccb6887-a207-45df-a993-d7e65238b9a8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.940228 4818 generic.go:334] "Generic (PLEG): container finished" podID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerID="f56727a2c66c4ee4e1ab1f8f696efa618028690b75d1b5097835730b77425bcb" exitCode=0 Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.940340 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ccb6887-a207-45df-a993-d7e65238b9a8","Type":"ContainerDied","Data":"f56727a2c66c4ee4e1ab1f8f696efa618028690b75d1b5097835730b77425bcb"} Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.940582 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ccb6887-a207-45df-a993-d7e65238b9a8","Type":"ContainerDied","Data":"784833c48527f8c5b0853ee417f8aef628a9820eadf727dca101d0833e3ade06"} Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.940386 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.940610 4818 scope.go:117] "RemoveContainer" containerID="82d4f1d7f39e525e1e592f9eb128d604910b2e24aa3a7da1fe38631d379d11a6" Dec 03 06:49:18 crc kubenswrapper[4818]: I1203 06:49:18.993314 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.009298 4818 scope.go:117] "RemoveContainer" containerID="695ef63c2a501763ef29bb40dd5bb218b57bede54065e47e786e9872b759af94" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.012980 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.025003 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:49:19 crc kubenswrapper[4818]: E1203 06:49:19.025544 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerName="sg-core" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.025574 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerName="sg-core" Dec 03 06:49:19 crc kubenswrapper[4818]: E1203 06:49:19.025596 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerName="ceilometer-notification-agent" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.025608 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerName="ceilometer-notification-agent" Dec 03 06:49:19 crc kubenswrapper[4818]: E1203 06:49:19.025636 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerName="ceilometer-central-agent" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.025648 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerName="ceilometer-central-agent" Dec 03 06:49:19 crc kubenswrapper[4818]: E1203 06:49:19.025683 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerName="proxy-httpd" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.025693 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerName="proxy-httpd" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.025950 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerName="ceilometer-notification-agent" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.025969 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerName="sg-core" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.025982 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerName="ceilometer-central-agent" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.026004 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" containerName="proxy-httpd" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.028036 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.031326 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.033178 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.033280 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.035958 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.041257 4818 scope.go:117] "RemoveContainer" containerID="f56727a2c66c4ee4e1ab1f8f696efa618028690b75d1b5097835730b77425bcb" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.076098 4818 scope.go:117] "RemoveContainer" containerID="b35ca1ef7881533da7ef2a5095163390a08d73582c41b819e855bcb5a4bf71e9" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.103341 4818 scope.go:117] "RemoveContainer" containerID="82d4f1d7f39e525e1e592f9eb128d604910b2e24aa3a7da1fe38631d379d11a6" Dec 03 06:49:19 crc kubenswrapper[4818]: E1203 06:49:19.103885 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82d4f1d7f39e525e1e592f9eb128d604910b2e24aa3a7da1fe38631d379d11a6\": container with ID starting with 82d4f1d7f39e525e1e592f9eb128d604910b2e24aa3a7da1fe38631d379d11a6 not found: ID does not exist" containerID="82d4f1d7f39e525e1e592f9eb128d604910b2e24aa3a7da1fe38631d379d11a6" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.103937 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82d4f1d7f39e525e1e592f9eb128d604910b2e24aa3a7da1fe38631d379d11a6"} err="failed to get container status \"82d4f1d7f39e525e1e592f9eb128d604910b2e24aa3a7da1fe38631d379d11a6\": rpc error: code = NotFound desc = could not find container \"82d4f1d7f39e525e1e592f9eb128d604910b2e24aa3a7da1fe38631d379d11a6\": container with ID starting with 82d4f1d7f39e525e1e592f9eb128d604910b2e24aa3a7da1fe38631d379d11a6 not found: ID does not exist" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.103971 4818 scope.go:117] "RemoveContainer" containerID="695ef63c2a501763ef29bb40dd5bb218b57bede54065e47e786e9872b759af94" Dec 03 06:49:19 crc kubenswrapper[4818]: E1203 06:49:19.104462 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"695ef63c2a501763ef29bb40dd5bb218b57bede54065e47e786e9872b759af94\": container with ID starting with 695ef63c2a501763ef29bb40dd5bb218b57bede54065e47e786e9872b759af94 not found: ID does not exist" containerID="695ef63c2a501763ef29bb40dd5bb218b57bede54065e47e786e9872b759af94" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.104486 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"695ef63c2a501763ef29bb40dd5bb218b57bede54065e47e786e9872b759af94"} err="failed to get container status \"695ef63c2a501763ef29bb40dd5bb218b57bede54065e47e786e9872b759af94\": rpc error: code = NotFound desc = could not find container \"695ef63c2a501763ef29bb40dd5bb218b57bede54065e47e786e9872b759af94\": container with ID starting with 695ef63c2a501763ef29bb40dd5bb218b57bede54065e47e786e9872b759af94 not found: ID does not exist" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.104501 4818 scope.go:117] "RemoveContainer" containerID="f56727a2c66c4ee4e1ab1f8f696efa618028690b75d1b5097835730b77425bcb" Dec 03 06:49:19 crc kubenswrapper[4818]: E1203 06:49:19.104892 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f56727a2c66c4ee4e1ab1f8f696efa618028690b75d1b5097835730b77425bcb\": container with ID starting with f56727a2c66c4ee4e1ab1f8f696efa618028690b75d1b5097835730b77425bcb not found: ID does not exist" containerID="f56727a2c66c4ee4e1ab1f8f696efa618028690b75d1b5097835730b77425bcb" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.104931 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f56727a2c66c4ee4e1ab1f8f696efa618028690b75d1b5097835730b77425bcb"} err="failed to get container status \"f56727a2c66c4ee4e1ab1f8f696efa618028690b75d1b5097835730b77425bcb\": rpc error: code = NotFound desc = could not find container \"f56727a2c66c4ee4e1ab1f8f696efa618028690b75d1b5097835730b77425bcb\": container with ID starting with f56727a2c66c4ee4e1ab1f8f696efa618028690b75d1b5097835730b77425bcb not found: ID does not exist" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.104963 4818 scope.go:117] "RemoveContainer" containerID="b35ca1ef7881533da7ef2a5095163390a08d73582c41b819e855bcb5a4bf71e9" Dec 03 06:49:19 crc kubenswrapper[4818]: E1203 06:49:19.105261 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b35ca1ef7881533da7ef2a5095163390a08d73582c41b819e855bcb5a4bf71e9\": container with ID starting with b35ca1ef7881533da7ef2a5095163390a08d73582c41b819e855bcb5a4bf71e9 not found: ID does not exist" containerID="b35ca1ef7881533da7ef2a5095163390a08d73582c41b819e855bcb5a4bf71e9" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.105286 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b35ca1ef7881533da7ef2a5095163390a08d73582c41b819e855bcb5a4bf71e9"} err="failed to get container status \"b35ca1ef7881533da7ef2a5095163390a08d73582c41b819e855bcb5a4bf71e9\": rpc error: code = NotFound desc = could not find container \"b35ca1ef7881533da7ef2a5095163390a08d73582c41b819e855bcb5a4bf71e9\": container with ID starting with b35ca1ef7881533da7ef2a5095163390a08d73582c41b819e855bcb5a4bf71e9 not found: ID does not exist" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.113046 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.113118 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.113256 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-config-data\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.113288 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8xnr\" (UniqueName: \"kubernetes.io/projected/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-kube-api-access-w8xnr\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.113332 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-run-httpd\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.113359 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-log-httpd\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.113555 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-scripts\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.113673 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.216342 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-scripts\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.216469 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.216621 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.216721 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.216807 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-config-data\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.216899 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8xnr\" (UniqueName: \"kubernetes.io/projected/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-kube-api-access-w8xnr\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.216962 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-run-httpd\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.216984 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-log-httpd\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.220423 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-run-httpd\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.220502 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-log-httpd\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.221710 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-scripts\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.222305 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.224215 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.228753 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-config-data\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.240965 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.246565 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8xnr\" (UniqueName: \"kubernetes.io/projected/631b9437-fe40-43b7-b8ad-34b4cad6ffa1-kube-api-access-w8xnr\") pod \"ceilometer-0\" (UID: \"631b9437-fe40-43b7-b8ad-34b4cad6ffa1\") " pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.357304 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.479915 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.522955 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fz6zn\" (UniqueName: \"kubernetes.io/projected/7d412fb3-cb4c-406e-a88c-c7b4787443fd-kube-api-access-fz6zn\") pod \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\" (UID: \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\") " Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.523019 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d412fb3-cb4c-406e-a88c-c7b4787443fd-logs\") pod \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\" (UID: \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\") " Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.523343 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d412fb3-cb4c-406e-a88c-c7b4787443fd-combined-ca-bundle\") pod \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\" (UID: \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\") " Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.523382 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d412fb3-cb4c-406e-a88c-c7b4787443fd-config-data\") pod \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\" (UID: \"7d412fb3-cb4c-406e-a88c-c7b4787443fd\") " Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.523651 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d412fb3-cb4c-406e-a88c-c7b4787443fd-logs" (OuterVolumeSpecName: "logs") pod "7d412fb3-cb4c-406e-a88c-c7b4787443fd" (UID: "7d412fb3-cb4c-406e-a88c-c7b4787443fd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.524215 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d412fb3-cb4c-406e-a88c-c7b4787443fd-logs\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.527663 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d412fb3-cb4c-406e-a88c-c7b4787443fd-kube-api-access-fz6zn" (OuterVolumeSpecName: "kube-api-access-fz6zn") pod "7d412fb3-cb4c-406e-a88c-c7b4787443fd" (UID: "7d412fb3-cb4c-406e-a88c-c7b4787443fd"). InnerVolumeSpecName "kube-api-access-fz6zn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.570720 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d412fb3-cb4c-406e-a88c-c7b4787443fd-config-data" (OuterVolumeSpecName: "config-data") pod "7d412fb3-cb4c-406e-a88c-c7b4787443fd" (UID: "7d412fb3-cb4c-406e-a88c-c7b4787443fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.574977 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d412fb3-cb4c-406e-a88c-c7b4787443fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d412fb3-cb4c-406e-a88c-c7b4787443fd" (UID: "7d412fb3-cb4c-406e-a88c-c7b4787443fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.625827 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d412fb3-cb4c-406e-a88c-c7b4787443fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.625858 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d412fb3-cb4c-406e-a88c-c7b4787443fd-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.625870 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fz6zn\" (UniqueName: \"kubernetes.io/projected/7d412fb3-cb4c-406e-a88c-c7b4787443fd-kube-api-access-fz6zn\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.853927 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 06:49:19 crc kubenswrapper[4818]: W1203 06:49:19.856333 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod631b9437_fe40_43b7_b8ad_34b4cad6ffa1.slice/crio-f49d74051c959ab52e0e469173e1c51296b9084e3fbaf7e997d66d3840a9068d WatchSource:0}: Error finding container f49d74051c959ab52e0e469173e1c51296b9084e3fbaf7e997d66d3840a9068d: Status 404 returned error can't find the container with id f49d74051c959ab52e0e469173e1c51296b9084e3fbaf7e997d66d3840a9068d Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.859497 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.959627 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"631b9437-fe40-43b7-b8ad-34b4cad6ffa1","Type":"ContainerStarted","Data":"f49d74051c959ab52e0e469173e1c51296b9084e3fbaf7e997d66d3840a9068d"} Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.962247 4818 generic.go:334] "Generic (PLEG): container finished" podID="7d412fb3-cb4c-406e-a88c-c7b4787443fd" containerID="b8cd7ec7d56d8f999b3b911e243bd10aaffdadae0514351f52d72a16606ddabd" exitCode=0 Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.962282 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d412fb3-cb4c-406e-a88c-c7b4787443fd","Type":"ContainerDied","Data":"b8cd7ec7d56d8f999b3b911e243bd10aaffdadae0514351f52d72a16606ddabd"} Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.962317 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7d412fb3-cb4c-406e-a88c-c7b4787443fd","Type":"ContainerDied","Data":"0d5ed0331e10c84b13b4ae807748d391201ad32c09afbea90a149e3b104ebd23"} Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.962333 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.962338 4818 scope.go:117] "RemoveContainer" containerID="b8cd7ec7d56d8f999b3b911e243bd10aaffdadae0514351f52d72a16606ddabd" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.997395 4818 scope.go:117] "RemoveContainer" containerID="af04fa69b1fb083430371c337c587687d7272cdd989650451cba86405d5884b6" Dec 03 06:49:19 crc kubenswrapper[4818]: I1203 06:49:19.997477 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.024619 4818 scope.go:117] "RemoveContainer" containerID="b8cd7ec7d56d8f999b3b911e243bd10aaffdadae0514351f52d72a16606ddabd" Dec 03 06:49:20 crc kubenswrapper[4818]: E1203 06:49:20.031396 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8cd7ec7d56d8f999b3b911e243bd10aaffdadae0514351f52d72a16606ddabd\": container with ID starting with b8cd7ec7d56d8f999b3b911e243bd10aaffdadae0514351f52d72a16606ddabd not found: ID does not exist" containerID="b8cd7ec7d56d8f999b3b911e243bd10aaffdadae0514351f52d72a16606ddabd" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.031438 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8cd7ec7d56d8f999b3b911e243bd10aaffdadae0514351f52d72a16606ddabd"} err="failed to get container status \"b8cd7ec7d56d8f999b3b911e243bd10aaffdadae0514351f52d72a16606ddabd\": rpc error: code = NotFound desc = could not find container \"b8cd7ec7d56d8f999b3b911e243bd10aaffdadae0514351f52d72a16606ddabd\": container with ID starting with b8cd7ec7d56d8f999b3b911e243bd10aaffdadae0514351f52d72a16606ddabd not found: ID does not exist" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.031462 4818 scope.go:117] "RemoveContainer" containerID="af04fa69b1fb083430371c337c587687d7272cdd989650451cba86405d5884b6" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.031764 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 06:49:20 crc kubenswrapper[4818]: E1203 06:49:20.032274 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af04fa69b1fb083430371c337c587687d7272cdd989650451cba86405d5884b6\": container with ID starting with af04fa69b1fb083430371c337c587687d7272cdd989650451cba86405d5884b6 not found: ID does not exist" containerID="af04fa69b1fb083430371c337c587687d7272cdd989650451cba86405d5884b6" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.032311 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af04fa69b1fb083430371c337c587687d7272cdd989650451cba86405d5884b6"} err="failed to get container status \"af04fa69b1fb083430371c337c587687d7272cdd989650451cba86405d5884b6\": rpc error: code = NotFound desc = could not find container \"af04fa69b1fb083430371c337c587687d7272cdd989650451cba86405d5884b6\": container with ID starting with af04fa69b1fb083430371c337c587687d7272cdd989650451cba86405d5884b6 not found: ID does not exist" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.036988 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 06:49:20 crc kubenswrapper[4818]: E1203 06:49:20.037462 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d412fb3-cb4c-406e-a88c-c7b4787443fd" containerName="nova-api-api" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.037489 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d412fb3-cb4c-406e-a88c-c7b4787443fd" containerName="nova-api-api" Dec 03 06:49:20 crc kubenswrapper[4818]: E1203 06:49:20.037525 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d412fb3-cb4c-406e-a88c-c7b4787443fd" containerName="nova-api-log" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.037536 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d412fb3-cb4c-406e-a88c-c7b4787443fd" containerName="nova-api-log" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.037738 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d412fb3-cb4c-406e-a88c-c7b4787443fd" containerName="nova-api-log" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.037760 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d412fb3-cb4c-406e-a88c-c7b4787443fd" containerName="nova-api-api" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.038853 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.041273 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.041310 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.041436 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.048707 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.136445 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.136912 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e35d84fc-186b-42b5-affa-2774576a0602-logs\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.137050 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.137104 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llgqb\" (UniqueName: \"kubernetes.io/projected/e35d84fc-186b-42b5-affa-2774576a0602-kube-api-access-llgqb\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.137132 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-config-data\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.137157 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-public-tls-certs\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.238361 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llgqb\" (UniqueName: \"kubernetes.io/projected/e35d84fc-186b-42b5-affa-2774576a0602-kube-api-access-llgqb\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.238426 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-config-data\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.238461 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-public-tls-certs\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.238522 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.238648 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e35d84fc-186b-42b5-affa-2774576a0602-logs\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.238689 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.239747 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e35d84fc-186b-42b5-affa-2774576a0602-logs\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.244720 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.246312 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-public-tls-certs\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.246505 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.249506 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-config-data\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.262308 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llgqb\" (UniqueName: \"kubernetes.io/projected/e35d84fc-186b-42b5-affa-2774576a0602-kube-api-access-llgqb\") pod \"nova-api-0\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.371831 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.760858 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ccb6887-a207-45df-a993-d7e65238b9a8" path="/var/lib/kubelet/pods/4ccb6887-a207-45df-a993-d7e65238b9a8/volumes" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.762358 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d412fb3-cb4c-406e-a88c-c7b4787443fd" path="/var/lib/kubelet/pods/7d412fb3-cb4c-406e-a88c-c7b4787443fd/volumes" Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.855104 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 06:49:20 crc kubenswrapper[4818]: W1203 06:49:20.857335 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode35d84fc_186b_42b5_affa_2774576a0602.slice/crio-f34b47be646ae5e37974ca7de09358cd6ef2f48359c6b8d2860480aca7f13f2a WatchSource:0}: Error finding container f34b47be646ae5e37974ca7de09358cd6ef2f48359c6b8d2860480aca7f13f2a: Status 404 returned error can't find the container with id f34b47be646ae5e37974ca7de09358cd6ef2f48359c6b8d2860480aca7f13f2a Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.973493 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e35d84fc-186b-42b5-affa-2774576a0602","Type":"ContainerStarted","Data":"f34b47be646ae5e37974ca7de09358cd6ef2f48359c6b8d2860480aca7f13f2a"} Dec 03 06:49:20 crc kubenswrapper[4818]: I1203 06:49:20.976582 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"631b9437-fe40-43b7-b8ad-34b4cad6ffa1","Type":"ContainerStarted","Data":"f9a7d48862c6d47ccebfe010135edd57b4cc995ab922f6635246fce1efa41f40"} Dec 03 06:49:21 crc kubenswrapper[4818]: I1203 06:49:21.250972 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:21 crc kubenswrapper[4818]: I1203 06:49:21.276085 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:21 crc kubenswrapper[4818]: I1203 06:49:21.988578 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e35d84fc-186b-42b5-affa-2774576a0602","Type":"ContainerStarted","Data":"343622c5b398d9e5f63c0292d2bf48bd4f2f014f2269f2ab983c53b4a6c8f00c"} Dec 03 06:49:21 crc kubenswrapper[4818]: I1203 06:49:21.988639 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e35d84fc-186b-42b5-affa-2774576a0602","Type":"ContainerStarted","Data":"1e9998804f6eddb0e79b2288b408be35e63267e62b2f553b5d9770a31c00a6f8"} Dec 03 06:49:21 crc kubenswrapper[4818]: I1203 06:49:21.991857 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"631b9437-fe40-43b7-b8ad-34b4cad6ffa1","Type":"ContainerStarted","Data":"17e0fb868450200d31570164dd3dc444c8fc846a2409fef890d1ff557a319672"} Dec 03 06:49:21 crc kubenswrapper[4818]: I1203 06:49:21.991954 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"631b9437-fe40-43b7-b8ad-34b4cad6ffa1","Type":"ContainerStarted","Data":"f8ded18656e031869e29cdfd20037f86ba133b02736cab453562fcb4a8a4a73f"} Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.006511 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.032324 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.03227272 podStartE2EDuration="3.03227272s" podCreationTimestamp="2025-12-03 06:49:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:49:22.016680484 +0000 UTC m=+1319.708289236" watchObservedRunningTime="2025-12-03 06:49:22.03227272 +0000 UTC m=+1319.723881492" Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.226971 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-flwjh"] Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.228575 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-flwjh" Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.230927 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.231149 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.247013 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-flwjh"] Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.285947 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f66g5\" (UniqueName: \"kubernetes.io/projected/e23ee654-89e4-449e-8417-23775d4dcbfb-kube-api-access-f66g5\") pod \"nova-cell1-cell-mapping-flwjh\" (UID: \"e23ee654-89e4-449e-8417-23775d4dcbfb\") " pod="openstack/nova-cell1-cell-mapping-flwjh" Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.286012 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-flwjh\" (UID: \"e23ee654-89e4-449e-8417-23775d4dcbfb\") " pod="openstack/nova-cell1-cell-mapping-flwjh" Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.286035 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-scripts\") pod \"nova-cell1-cell-mapping-flwjh\" (UID: \"e23ee654-89e4-449e-8417-23775d4dcbfb\") " pod="openstack/nova-cell1-cell-mapping-flwjh" Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.286115 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-config-data\") pod \"nova-cell1-cell-mapping-flwjh\" (UID: \"e23ee654-89e4-449e-8417-23775d4dcbfb\") " pod="openstack/nova-cell1-cell-mapping-flwjh" Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.388560 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f66g5\" (UniqueName: \"kubernetes.io/projected/e23ee654-89e4-449e-8417-23775d4dcbfb-kube-api-access-f66g5\") pod \"nova-cell1-cell-mapping-flwjh\" (UID: \"e23ee654-89e4-449e-8417-23775d4dcbfb\") " pod="openstack/nova-cell1-cell-mapping-flwjh" Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.388636 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-flwjh\" (UID: \"e23ee654-89e4-449e-8417-23775d4dcbfb\") " pod="openstack/nova-cell1-cell-mapping-flwjh" Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.388657 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-scripts\") pod \"nova-cell1-cell-mapping-flwjh\" (UID: \"e23ee654-89e4-449e-8417-23775d4dcbfb\") " pod="openstack/nova-cell1-cell-mapping-flwjh" Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.388708 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-config-data\") pod \"nova-cell1-cell-mapping-flwjh\" (UID: \"e23ee654-89e4-449e-8417-23775d4dcbfb\") " pod="openstack/nova-cell1-cell-mapping-flwjh" Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.396705 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-scripts\") pod \"nova-cell1-cell-mapping-flwjh\" (UID: \"e23ee654-89e4-449e-8417-23775d4dcbfb\") " pod="openstack/nova-cell1-cell-mapping-flwjh" Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.396726 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-flwjh\" (UID: \"e23ee654-89e4-449e-8417-23775d4dcbfb\") " pod="openstack/nova-cell1-cell-mapping-flwjh" Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.396905 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-config-data\") pod \"nova-cell1-cell-mapping-flwjh\" (UID: \"e23ee654-89e4-449e-8417-23775d4dcbfb\") " pod="openstack/nova-cell1-cell-mapping-flwjh" Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.406746 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f66g5\" (UniqueName: \"kubernetes.io/projected/e23ee654-89e4-449e-8417-23775d4dcbfb-kube-api-access-f66g5\") pod \"nova-cell1-cell-mapping-flwjh\" (UID: \"e23ee654-89e4-449e-8417-23775d4dcbfb\") " pod="openstack/nova-cell1-cell-mapping-flwjh" Dec 03 06:49:22 crc kubenswrapper[4818]: I1203 06:49:22.562695 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-flwjh" Dec 03 06:49:23 crc kubenswrapper[4818]: W1203 06:49:23.025222 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode23ee654_89e4_449e_8417_23775d4dcbfb.slice/crio-7f6cfe0f253a48831687840eb943630bcc3c1d2a21c6e667bd5f9bee818f3fd7 WatchSource:0}: Error finding container 7f6cfe0f253a48831687840eb943630bcc3c1d2a21c6e667bd5f9bee818f3fd7: Status 404 returned error can't find the container with id 7f6cfe0f253a48831687840eb943630bcc3c1d2a21c6e667bd5f9bee818f3fd7 Dec 03 06:49:23 crc kubenswrapper[4818]: I1203 06:49:23.036890 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-flwjh"] Dec 03 06:49:23 crc kubenswrapper[4818]: I1203 06:49:23.438421 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:49:23 crc kubenswrapper[4818]: I1203 06:49:23.501422 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-tzstk"] Dec 03 06:49:23 crc kubenswrapper[4818]: I1203 06:49:23.501660 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-tzstk" podUID="4882a5a6-4fb7-4b00-b8bf-c28e116b5c16" containerName="dnsmasq-dns" containerID="cri-o://ed1281a3e82bb1946ebfd13ddb73195dbfa3219ba39b8229231fbecab8c9c184" gracePeriod=10 Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.008799 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.010478 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-flwjh" event={"ID":"e23ee654-89e4-449e-8417-23775d4dcbfb","Type":"ContainerStarted","Data":"ffbb9c6e2bf23ddfd97de203be0f20b6bd929349b8383901240c6a2c7f15a828"} Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.010590 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-flwjh" event={"ID":"e23ee654-89e4-449e-8417-23775d4dcbfb","Type":"ContainerStarted","Data":"7f6cfe0f253a48831687840eb943630bcc3c1d2a21c6e667bd5f9bee818f3fd7"} Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.013263 4818 generic.go:334] "Generic (PLEG): container finished" podID="4882a5a6-4fb7-4b00-b8bf-c28e116b5c16" containerID="ed1281a3e82bb1946ebfd13ddb73195dbfa3219ba39b8229231fbecab8c9c184" exitCode=0 Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.013318 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-tzstk" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.013336 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-tzstk" event={"ID":"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16","Type":"ContainerDied","Data":"ed1281a3e82bb1946ebfd13ddb73195dbfa3219ba39b8229231fbecab8c9c184"} Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.013450 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-tzstk" event={"ID":"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16","Type":"ContainerDied","Data":"507b2612e1aa77e754128452015e5d4264779d3ac2fd661f2c02f8f58ece9a41"} Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.013473 4818 scope.go:117] "RemoveContainer" containerID="ed1281a3e82bb1946ebfd13ddb73195dbfa3219ba39b8229231fbecab8c9c184" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.016043 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"631b9437-fe40-43b7-b8ad-34b4cad6ffa1","Type":"ContainerStarted","Data":"7169b6b025292a957709a345c075c6ec3b993b56637bfdea9a6d3dd12e4cb1fa"} Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.023204 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.027574 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-ovsdbserver-sb\") pod \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.028449 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-dns-swift-storage-0\") pod \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.028546 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2s6mg\" (UniqueName: \"kubernetes.io/projected/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-kube-api-access-2s6mg\") pod \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.028669 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-config\") pod \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.028763 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-ovsdbserver-nb\") pod \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.028938 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-dns-svc\") pod \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.036478 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-kube-api-access-2s6mg" (OuterVolumeSpecName: "kube-api-access-2s6mg") pod "4882a5a6-4fb7-4b00-b8bf-c28e116b5c16" (UID: "4882a5a6-4fb7-4b00-b8bf-c28e116b5c16"). InnerVolumeSpecName "kube-api-access-2s6mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.052966 4818 scope.go:117] "RemoveContainer" containerID="395e1d492ca211df82ce93315e1fac07e8eda09362461138bb52fcf64d81198e" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.078768 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.953677818 podStartE2EDuration="6.078745756s" podCreationTimestamp="2025-12-03 06:49:18 +0000 UTC" firstStartedPulling="2025-12-03 06:49:19.859181645 +0000 UTC m=+1317.550790397" lastFinishedPulling="2025-12-03 06:49:22.984249583 +0000 UTC m=+1320.675858335" observedRunningTime="2025-12-03 06:49:24.061614482 +0000 UTC m=+1321.753223234" watchObservedRunningTime="2025-12-03 06:49:24.078745756 +0000 UTC m=+1321.770354508" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.129647 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-flwjh" podStartSLOduration=2.129620063 podStartE2EDuration="2.129620063s" podCreationTimestamp="2025-12-03 06:49:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:49:24.088921577 +0000 UTC m=+1321.780530349" watchObservedRunningTime="2025-12-03 06:49:24.129620063 +0000 UTC m=+1321.821228825" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.131094 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4882a5a6-4fb7-4b00-b8bf-c28e116b5c16" (UID: "4882a5a6-4fb7-4b00-b8bf-c28e116b5c16"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.131377 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4882a5a6-4fb7-4b00-b8bf-c28e116b5c16" (UID: "4882a5a6-4fb7-4b00-b8bf-c28e116b5c16"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.131520 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-dns-swift-storage-0\") pod \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\" (UID: \"4882a5a6-4fb7-4b00-b8bf-c28e116b5c16\") " Dec 03 06:49:24 crc kubenswrapper[4818]: W1203 06:49:24.133591 4818 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16/volumes/kubernetes.io~configmap/dns-swift-storage-0 Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.133621 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4882a5a6-4fb7-4b00-b8bf-c28e116b5c16" (UID: "4882a5a6-4fb7-4b00-b8bf-c28e116b5c16"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.136502 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.136536 4818 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.136549 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2s6mg\" (UniqueName: \"kubernetes.io/projected/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-kube-api-access-2s6mg\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.141874 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-config" (OuterVolumeSpecName: "config") pod "4882a5a6-4fb7-4b00-b8bf-c28e116b5c16" (UID: "4882a5a6-4fb7-4b00-b8bf-c28e116b5c16"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.142571 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4882a5a6-4fb7-4b00-b8bf-c28e116b5c16" (UID: "4882a5a6-4fb7-4b00-b8bf-c28e116b5c16"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.147311 4818 scope.go:117] "RemoveContainer" containerID="ed1281a3e82bb1946ebfd13ddb73195dbfa3219ba39b8229231fbecab8c9c184" Dec 03 06:49:24 crc kubenswrapper[4818]: E1203 06:49:24.148560 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed1281a3e82bb1946ebfd13ddb73195dbfa3219ba39b8229231fbecab8c9c184\": container with ID starting with ed1281a3e82bb1946ebfd13ddb73195dbfa3219ba39b8229231fbecab8c9c184 not found: ID does not exist" containerID="ed1281a3e82bb1946ebfd13ddb73195dbfa3219ba39b8229231fbecab8c9c184" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.148589 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed1281a3e82bb1946ebfd13ddb73195dbfa3219ba39b8229231fbecab8c9c184"} err="failed to get container status \"ed1281a3e82bb1946ebfd13ddb73195dbfa3219ba39b8229231fbecab8c9c184\": rpc error: code = NotFound desc = could not find container \"ed1281a3e82bb1946ebfd13ddb73195dbfa3219ba39b8229231fbecab8c9c184\": container with ID starting with ed1281a3e82bb1946ebfd13ddb73195dbfa3219ba39b8229231fbecab8c9c184 not found: ID does not exist" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.148606 4818 scope.go:117] "RemoveContainer" containerID="395e1d492ca211df82ce93315e1fac07e8eda09362461138bb52fcf64d81198e" Dec 03 06:49:24 crc kubenswrapper[4818]: E1203 06:49:24.149057 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"395e1d492ca211df82ce93315e1fac07e8eda09362461138bb52fcf64d81198e\": container with ID starting with 395e1d492ca211df82ce93315e1fac07e8eda09362461138bb52fcf64d81198e not found: ID does not exist" containerID="395e1d492ca211df82ce93315e1fac07e8eda09362461138bb52fcf64d81198e" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.149099 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"395e1d492ca211df82ce93315e1fac07e8eda09362461138bb52fcf64d81198e"} err="failed to get container status \"395e1d492ca211df82ce93315e1fac07e8eda09362461138bb52fcf64d81198e\": rpc error: code = NotFound desc = could not find container \"395e1d492ca211df82ce93315e1fac07e8eda09362461138bb52fcf64d81198e\": container with ID starting with 395e1d492ca211df82ce93315e1fac07e8eda09362461138bb52fcf64d81198e not found: ID does not exist" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.164472 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4882a5a6-4fb7-4b00-b8bf-c28e116b5c16" (UID: "4882a5a6-4fb7-4b00-b8bf-c28e116b5c16"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.238065 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.238104 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.238116 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.346343 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-tzstk"] Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.356086 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-tzstk"] Dec 03 06:49:24 crc kubenswrapper[4818]: I1203 06:49:24.748766 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4882a5a6-4fb7-4b00-b8bf-c28e116b5c16" path="/var/lib/kubelet/pods/4882a5a6-4fb7-4b00-b8bf-c28e116b5c16/volumes" Dec 03 06:49:28 crc kubenswrapper[4818]: I1203 06:49:28.057186 4818 generic.go:334] "Generic (PLEG): container finished" podID="e23ee654-89e4-449e-8417-23775d4dcbfb" containerID="ffbb9c6e2bf23ddfd97de203be0f20b6bd929349b8383901240c6a2c7f15a828" exitCode=0 Dec 03 06:49:28 crc kubenswrapper[4818]: I1203 06:49:28.057250 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-flwjh" event={"ID":"e23ee654-89e4-449e-8417-23775d4dcbfb","Type":"ContainerDied","Data":"ffbb9c6e2bf23ddfd97de203be0f20b6bd929349b8383901240c6a2c7f15a828"} Dec 03 06:49:29 crc kubenswrapper[4818]: I1203 06:49:29.483956 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-flwjh" Dec 03 06:49:29 crc kubenswrapper[4818]: I1203 06:49:29.652467 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-config-data\") pod \"e23ee654-89e4-449e-8417-23775d4dcbfb\" (UID: \"e23ee654-89e4-449e-8417-23775d4dcbfb\") " Dec 03 06:49:29 crc kubenswrapper[4818]: I1203 06:49:29.652774 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f66g5\" (UniqueName: \"kubernetes.io/projected/e23ee654-89e4-449e-8417-23775d4dcbfb-kube-api-access-f66g5\") pod \"e23ee654-89e4-449e-8417-23775d4dcbfb\" (UID: \"e23ee654-89e4-449e-8417-23775d4dcbfb\") " Dec 03 06:49:29 crc kubenswrapper[4818]: I1203 06:49:29.653072 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-combined-ca-bundle\") pod \"e23ee654-89e4-449e-8417-23775d4dcbfb\" (UID: \"e23ee654-89e4-449e-8417-23775d4dcbfb\") " Dec 03 06:49:29 crc kubenswrapper[4818]: I1203 06:49:29.654321 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-scripts\") pod \"e23ee654-89e4-449e-8417-23775d4dcbfb\" (UID: \"e23ee654-89e4-449e-8417-23775d4dcbfb\") " Dec 03 06:49:29 crc kubenswrapper[4818]: I1203 06:49:29.663889 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e23ee654-89e4-449e-8417-23775d4dcbfb-kube-api-access-f66g5" (OuterVolumeSpecName: "kube-api-access-f66g5") pod "e23ee654-89e4-449e-8417-23775d4dcbfb" (UID: "e23ee654-89e4-449e-8417-23775d4dcbfb"). InnerVolumeSpecName "kube-api-access-f66g5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:49:29 crc kubenswrapper[4818]: I1203 06:49:29.664367 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-scripts" (OuterVolumeSpecName: "scripts") pod "e23ee654-89e4-449e-8417-23775d4dcbfb" (UID: "e23ee654-89e4-449e-8417-23775d4dcbfb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:29 crc kubenswrapper[4818]: I1203 06:49:29.681650 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e23ee654-89e4-449e-8417-23775d4dcbfb" (UID: "e23ee654-89e4-449e-8417-23775d4dcbfb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:29 crc kubenswrapper[4818]: I1203 06:49:29.682112 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-config-data" (OuterVolumeSpecName: "config-data") pod "e23ee654-89e4-449e-8417-23775d4dcbfb" (UID: "e23ee654-89e4-449e-8417-23775d4dcbfb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:29 crc kubenswrapper[4818]: I1203 06:49:29.758443 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:29 crc kubenswrapper[4818]: I1203 06:49:29.758502 4818 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:29 crc kubenswrapper[4818]: I1203 06:49:29.758527 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23ee654-89e4-449e-8417-23775d4dcbfb-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:29 crc kubenswrapper[4818]: I1203 06:49:29.758550 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f66g5\" (UniqueName: \"kubernetes.io/projected/e23ee654-89e4-449e-8417-23775d4dcbfb-kube-api-access-f66g5\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:30 crc kubenswrapper[4818]: I1203 06:49:30.084371 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-flwjh" event={"ID":"e23ee654-89e4-449e-8417-23775d4dcbfb","Type":"ContainerDied","Data":"7f6cfe0f253a48831687840eb943630bcc3c1d2a21c6e667bd5f9bee818f3fd7"} Dec 03 06:49:30 crc kubenswrapper[4818]: I1203 06:49:30.084434 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f6cfe0f253a48831687840eb943630bcc3c1d2a21c6e667bd5f9bee818f3fd7" Dec 03 06:49:30 crc kubenswrapper[4818]: I1203 06:49:30.084455 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-flwjh" Dec 03 06:49:30 crc kubenswrapper[4818]: I1203 06:49:30.271457 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 06:49:30 crc kubenswrapper[4818]: I1203 06:49:30.271747 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e35d84fc-186b-42b5-affa-2774576a0602" containerName="nova-api-log" containerID="cri-o://1e9998804f6eddb0e79b2288b408be35e63267e62b2f553b5d9770a31c00a6f8" gracePeriod=30 Dec 03 06:49:30 crc kubenswrapper[4818]: I1203 06:49:30.271774 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e35d84fc-186b-42b5-affa-2774576a0602" containerName="nova-api-api" containerID="cri-o://343622c5b398d9e5f63c0292d2bf48bd4f2f014f2269f2ab983c53b4a6c8f00c" gracePeriod=30 Dec 03 06:49:30 crc kubenswrapper[4818]: I1203 06:49:30.284973 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 06:49:30 crc kubenswrapper[4818]: I1203 06:49:30.285211 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="b87c588e-6d4c-49b9-87ac-1d99694ffafd" containerName="nova-scheduler-scheduler" containerID="cri-o://8511681a19b1f575d11ea59da1ebcd4ad4a52d0a90258354449e67f91d1e8f7d" gracePeriod=30 Dec 03 06:49:30 crc kubenswrapper[4818]: I1203 06:49:30.330610 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:49:30 crc kubenswrapper[4818]: I1203 06:49:30.331135 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8aa8deb3-c59d-4150-beb4-e68529020f60" containerName="nova-metadata-log" containerID="cri-o://af1f3345b5796f8d0b20c74bc648f214297ce4edb68bdd11ad027e991ae2237a" gracePeriod=30 Dec 03 06:49:30 crc kubenswrapper[4818]: I1203 06:49:30.331261 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8aa8deb3-c59d-4150-beb4-e68529020f60" containerName="nova-metadata-metadata" containerID="cri-o://beab1d8ca08727518b06bd399677f6f228fac5d0e7db7c6f66ec3fedc32caf63" gracePeriod=30 Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.114451 4818 generic.go:334] "Generic (PLEG): container finished" podID="e35d84fc-186b-42b5-affa-2774576a0602" containerID="343622c5b398d9e5f63c0292d2bf48bd4f2f014f2269f2ab983c53b4a6c8f00c" exitCode=0 Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.115501 4818 generic.go:334] "Generic (PLEG): container finished" podID="e35d84fc-186b-42b5-affa-2774576a0602" containerID="1e9998804f6eddb0e79b2288b408be35e63267e62b2f553b5d9770a31c00a6f8" exitCode=143 Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.114535 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e35d84fc-186b-42b5-affa-2774576a0602","Type":"ContainerDied","Data":"343622c5b398d9e5f63c0292d2bf48bd4f2f014f2269f2ab983c53b4a6c8f00c"} Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.115770 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e35d84fc-186b-42b5-affa-2774576a0602","Type":"ContainerDied","Data":"1e9998804f6eddb0e79b2288b408be35e63267e62b2f553b5d9770a31c00a6f8"} Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.117717 4818 generic.go:334] "Generic (PLEG): container finished" podID="8aa8deb3-c59d-4150-beb4-e68529020f60" containerID="af1f3345b5796f8d0b20c74bc648f214297ce4edb68bdd11ad027e991ae2237a" exitCode=143 Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.117802 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8aa8deb3-c59d-4150-beb4-e68529020f60","Type":"ContainerDied","Data":"af1f3345b5796f8d0b20c74bc648f214297ce4edb68bdd11ad027e991ae2237a"} Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.453664 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.617524 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-config-data\") pod \"e35d84fc-186b-42b5-affa-2774576a0602\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.617570 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-combined-ca-bundle\") pod \"e35d84fc-186b-42b5-affa-2774576a0602\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.617663 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e35d84fc-186b-42b5-affa-2774576a0602-logs\") pod \"e35d84fc-186b-42b5-affa-2774576a0602\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.617738 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-internal-tls-certs\") pod \"e35d84fc-186b-42b5-affa-2774576a0602\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.617805 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-public-tls-certs\") pod \"e35d84fc-186b-42b5-affa-2774576a0602\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.617858 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llgqb\" (UniqueName: \"kubernetes.io/projected/e35d84fc-186b-42b5-affa-2774576a0602-kube-api-access-llgqb\") pod \"e35d84fc-186b-42b5-affa-2774576a0602\" (UID: \"e35d84fc-186b-42b5-affa-2774576a0602\") " Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.619117 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e35d84fc-186b-42b5-affa-2774576a0602-logs" (OuterVolumeSpecName: "logs") pod "e35d84fc-186b-42b5-affa-2774576a0602" (UID: "e35d84fc-186b-42b5-affa-2774576a0602"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.622895 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e35d84fc-186b-42b5-affa-2774576a0602-kube-api-access-llgqb" (OuterVolumeSpecName: "kube-api-access-llgqb") pod "e35d84fc-186b-42b5-affa-2774576a0602" (UID: "e35d84fc-186b-42b5-affa-2774576a0602"). InnerVolumeSpecName "kube-api-access-llgqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.648088 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e35d84fc-186b-42b5-affa-2774576a0602" (UID: "e35d84fc-186b-42b5-affa-2774576a0602"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.654036 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-config-data" (OuterVolumeSpecName: "config-data") pod "e35d84fc-186b-42b5-affa-2774576a0602" (UID: "e35d84fc-186b-42b5-affa-2774576a0602"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.666800 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e35d84fc-186b-42b5-affa-2774576a0602" (UID: "e35d84fc-186b-42b5-affa-2774576a0602"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.674424 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e35d84fc-186b-42b5-affa-2774576a0602" (UID: "e35d84fc-186b-42b5-affa-2774576a0602"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.720054 4818 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.720090 4818 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.720103 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llgqb\" (UniqueName: \"kubernetes.io/projected/e35d84fc-186b-42b5-affa-2774576a0602-kube-api-access-llgqb\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.720118 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.720131 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e35d84fc-186b-42b5-affa-2774576a0602-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.720159 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e35d84fc-186b-42b5-affa-2774576a0602-logs\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.731488 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.923502 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzw6h\" (UniqueName: \"kubernetes.io/projected/b87c588e-6d4c-49b9-87ac-1d99694ffafd-kube-api-access-zzw6h\") pod \"b87c588e-6d4c-49b9-87ac-1d99694ffafd\" (UID: \"b87c588e-6d4c-49b9-87ac-1d99694ffafd\") " Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.923726 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b87c588e-6d4c-49b9-87ac-1d99694ffafd-config-data\") pod \"b87c588e-6d4c-49b9-87ac-1d99694ffafd\" (UID: \"b87c588e-6d4c-49b9-87ac-1d99694ffafd\") " Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.923971 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b87c588e-6d4c-49b9-87ac-1d99694ffafd-combined-ca-bundle\") pod \"b87c588e-6d4c-49b9-87ac-1d99694ffafd\" (UID: \"b87c588e-6d4c-49b9-87ac-1d99694ffafd\") " Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.934157 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b87c588e-6d4c-49b9-87ac-1d99694ffafd-kube-api-access-zzw6h" (OuterVolumeSpecName: "kube-api-access-zzw6h") pod "b87c588e-6d4c-49b9-87ac-1d99694ffafd" (UID: "b87c588e-6d4c-49b9-87ac-1d99694ffafd"). InnerVolumeSpecName "kube-api-access-zzw6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.953749 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b87c588e-6d4c-49b9-87ac-1d99694ffafd-config-data" (OuterVolumeSpecName: "config-data") pod "b87c588e-6d4c-49b9-87ac-1d99694ffafd" (UID: "b87c588e-6d4c-49b9-87ac-1d99694ffafd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:31 crc kubenswrapper[4818]: I1203 06:49:31.972391 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b87c588e-6d4c-49b9-87ac-1d99694ffafd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b87c588e-6d4c-49b9-87ac-1d99694ffafd" (UID: "b87c588e-6d4c-49b9-87ac-1d99694ffafd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.026740 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzw6h\" (UniqueName: \"kubernetes.io/projected/b87c588e-6d4c-49b9-87ac-1d99694ffafd-kube-api-access-zzw6h\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.026980 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b87c588e-6d4c-49b9-87ac-1d99694ffafd-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.027045 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b87c588e-6d4c-49b9-87ac-1d99694ffafd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.128365 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e35d84fc-186b-42b5-affa-2774576a0602","Type":"ContainerDied","Data":"f34b47be646ae5e37974ca7de09358cd6ef2f48359c6b8d2860480aca7f13f2a"} Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.128398 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.129249 4818 scope.go:117] "RemoveContainer" containerID="343622c5b398d9e5f63c0292d2bf48bd4f2f014f2269f2ab983c53b4a6c8f00c" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.129624 4818 generic.go:334] "Generic (PLEG): container finished" podID="b87c588e-6d4c-49b9-87ac-1d99694ffafd" containerID="8511681a19b1f575d11ea59da1ebcd4ad4a52d0a90258354449e67f91d1e8f7d" exitCode=0 Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.129721 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b87c588e-6d4c-49b9-87ac-1d99694ffafd","Type":"ContainerDied","Data":"8511681a19b1f575d11ea59da1ebcd4ad4a52d0a90258354449e67f91d1e8f7d"} Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.129789 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b87c588e-6d4c-49b9-87ac-1d99694ffafd","Type":"ContainerDied","Data":"6c004ef4e0ca44eb5aaa64697eb9a0a71c3ffced770a317931701a236a2dd5f3"} Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.129836 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.157352 4818 scope.go:117] "RemoveContainer" containerID="1e9998804f6eddb0e79b2288b408be35e63267e62b2f553b5d9770a31c00a6f8" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.188211 4818 scope.go:117] "RemoveContainer" containerID="8511681a19b1f575d11ea59da1ebcd4ad4a52d0a90258354449e67f91d1e8f7d" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.190872 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.203907 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.225035 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.240731 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.251042 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 06:49:32 crc kubenswrapper[4818]: E1203 06:49:32.251460 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b87c588e-6d4c-49b9-87ac-1d99694ffafd" containerName="nova-scheduler-scheduler" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.251478 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b87c588e-6d4c-49b9-87ac-1d99694ffafd" containerName="nova-scheduler-scheduler" Dec 03 06:49:32 crc kubenswrapper[4818]: E1203 06:49:32.251511 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4882a5a6-4fb7-4b00-b8bf-c28e116b5c16" containerName="dnsmasq-dns" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.251518 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4882a5a6-4fb7-4b00-b8bf-c28e116b5c16" containerName="dnsmasq-dns" Dec 03 06:49:32 crc kubenswrapper[4818]: E1203 06:49:32.251528 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e23ee654-89e4-449e-8417-23775d4dcbfb" containerName="nova-manage" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.251534 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e23ee654-89e4-449e-8417-23775d4dcbfb" containerName="nova-manage" Dec 03 06:49:32 crc kubenswrapper[4818]: E1203 06:49:32.251551 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4882a5a6-4fb7-4b00-b8bf-c28e116b5c16" containerName="init" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.251557 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4882a5a6-4fb7-4b00-b8bf-c28e116b5c16" containerName="init" Dec 03 06:49:32 crc kubenswrapper[4818]: E1203 06:49:32.251572 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e35d84fc-186b-42b5-affa-2774576a0602" containerName="nova-api-log" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.251577 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e35d84fc-186b-42b5-affa-2774576a0602" containerName="nova-api-log" Dec 03 06:49:32 crc kubenswrapper[4818]: E1203 06:49:32.251586 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e35d84fc-186b-42b5-affa-2774576a0602" containerName="nova-api-api" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.251591 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e35d84fc-186b-42b5-affa-2774576a0602" containerName="nova-api-api" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.251795 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4882a5a6-4fb7-4b00-b8bf-c28e116b5c16" containerName="dnsmasq-dns" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.251826 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="e23ee654-89e4-449e-8417-23775d4dcbfb" containerName="nova-manage" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.251837 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="e35d84fc-186b-42b5-affa-2774576a0602" containerName="nova-api-log" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.251853 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="b87c588e-6d4c-49b9-87ac-1d99694ffafd" containerName="nova-scheduler-scheduler" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.251866 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="e35d84fc-186b-42b5-affa-2774576a0602" containerName="nova-api-api" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.252054 4818 scope.go:117] "RemoveContainer" containerID="8511681a19b1f575d11ea59da1ebcd4ad4a52d0a90258354449e67f91d1e8f7d" Dec 03 06:49:32 crc kubenswrapper[4818]: E1203 06:49:32.252481 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8511681a19b1f575d11ea59da1ebcd4ad4a52d0a90258354449e67f91d1e8f7d\": container with ID starting with 8511681a19b1f575d11ea59da1ebcd4ad4a52d0a90258354449e67f91d1e8f7d not found: ID does not exist" containerID="8511681a19b1f575d11ea59da1ebcd4ad4a52d0a90258354449e67f91d1e8f7d" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.252510 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8511681a19b1f575d11ea59da1ebcd4ad4a52d0a90258354449e67f91d1e8f7d"} err="failed to get container status \"8511681a19b1f575d11ea59da1ebcd4ad4a52d0a90258354449e67f91d1e8f7d\": rpc error: code = NotFound desc = could not find container \"8511681a19b1f575d11ea59da1ebcd4ad4a52d0a90258354449e67f91d1e8f7d\": container with ID starting with 8511681a19b1f575d11ea59da1ebcd4ad4a52d0a90258354449e67f91d1e8f7d not found: ID does not exist" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.252831 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.256275 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.256430 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.256544 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.260864 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.262137 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.263743 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.277194 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.287898 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.435610 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8mcd\" (UniqueName: \"kubernetes.io/projected/5a5a4f96-9611-456a-a084-88ba9f3b548a-kube-api-access-k8mcd\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.436240 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a5a4f96-9611-456a-a084-88ba9f3b548a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.436353 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a5a4f96-9611-456a-a084-88ba9f3b548a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.436428 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a5a4f96-9611-456a-a084-88ba9f3b548a-logs\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.436598 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a5a4f96-9611-456a-a084-88ba9f3b548a-config-data\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.436714 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39a9a10c-e02f-4468-95a9-651b694b1f9a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"39a9a10c-e02f-4468-95a9-651b694b1f9a\") " pod="openstack/nova-scheduler-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.436785 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a5a4f96-9611-456a-a084-88ba9f3b548a-public-tls-certs\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.436900 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j85h5\" (UniqueName: \"kubernetes.io/projected/39a9a10c-e02f-4468-95a9-651b694b1f9a-kube-api-access-j85h5\") pod \"nova-scheduler-0\" (UID: \"39a9a10c-e02f-4468-95a9-651b694b1f9a\") " pod="openstack/nova-scheduler-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.436991 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39a9a10c-e02f-4468-95a9-651b694b1f9a-config-data\") pod \"nova-scheduler-0\" (UID: \"39a9a10c-e02f-4468-95a9-651b694b1f9a\") " pod="openstack/nova-scheduler-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.538329 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a5a4f96-9611-456a-a084-88ba9f3b548a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.538373 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a5a4f96-9611-456a-a084-88ba9f3b548a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.538394 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a5a4f96-9611-456a-a084-88ba9f3b548a-logs\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.538438 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a5a4f96-9611-456a-a084-88ba9f3b548a-config-data\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.538467 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39a9a10c-e02f-4468-95a9-651b694b1f9a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"39a9a10c-e02f-4468-95a9-651b694b1f9a\") " pod="openstack/nova-scheduler-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.538486 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a5a4f96-9611-456a-a084-88ba9f3b548a-public-tls-certs\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.538510 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j85h5\" (UniqueName: \"kubernetes.io/projected/39a9a10c-e02f-4468-95a9-651b694b1f9a-kube-api-access-j85h5\") pod \"nova-scheduler-0\" (UID: \"39a9a10c-e02f-4468-95a9-651b694b1f9a\") " pod="openstack/nova-scheduler-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.538533 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39a9a10c-e02f-4468-95a9-651b694b1f9a-config-data\") pod \"nova-scheduler-0\" (UID: \"39a9a10c-e02f-4468-95a9-651b694b1f9a\") " pod="openstack/nova-scheduler-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.538570 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8mcd\" (UniqueName: \"kubernetes.io/projected/5a5a4f96-9611-456a-a084-88ba9f3b548a-kube-api-access-k8mcd\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.539357 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a5a4f96-9611-456a-a084-88ba9f3b548a-logs\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.543782 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a5a4f96-9611-456a-a084-88ba9f3b548a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.544142 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a5a4f96-9611-456a-a084-88ba9f3b548a-config-data\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.545563 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39a9a10c-e02f-4468-95a9-651b694b1f9a-config-data\") pod \"nova-scheduler-0\" (UID: \"39a9a10c-e02f-4468-95a9-651b694b1f9a\") " pod="openstack/nova-scheduler-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.546314 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a5a4f96-9611-456a-a084-88ba9f3b548a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.547013 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a5a4f96-9611-456a-a084-88ba9f3b548a-public-tls-certs\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.550117 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39a9a10c-e02f-4468-95a9-651b694b1f9a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"39a9a10c-e02f-4468-95a9-651b694b1f9a\") " pod="openstack/nova-scheduler-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.568436 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8mcd\" (UniqueName: \"kubernetes.io/projected/5a5a4f96-9611-456a-a084-88ba9f3b548a-kube-api-access-k8mcd\") pod \"nova-api-0\" (UID: \"5a5a4f96-9611-456a-a084-88ba9f3b548a\") " pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.569734 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j85h5\" (UniqueName: \"kubernetes.io/projected/39a9a10c-e02f-4468-95a9-651b694b1f9a-kube-api-access-j85h5\") pod \"nova-scheduler-0\" (UID: \"39a9a10c-e02f-4468-95a9-651b694b1f9a\") " pod="openstack/nova-scheduler-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.581050 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.589526 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.761648 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b87c588e-6d4c-49b9-87ac-1d99694ffafd" path="/var/lib/kubelet/pods/b87c588e-6d4c-49b9-87ac-1d99694ffafd/volumes" Dec 03 06:49:32 crc kubenswrapper[4818]: I1203 06:49:32.762670 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e35d84fc-186b-42b5-affa-2774576a0602" path="/var/lib/kubelet/pods/e35d84fc-186b-42b5-affa-2774576a0602/volumes" Dec 03 06:49:33 crc kubenswrapper[4818]: I1203 06:49:33.100321 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 06:49:33 crc kubenswrapper[4818]: I1203 06:49:33.110144 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 06:49:33 crc kubenswrapper[4818]: I1203 06:49:33.141295 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"39a9a10c-e02f-4468-95a9-651b694b1f9a","Type":"ContainerStarted","Data":"0210a0868b04dd8a3532ab3be63d60cff8c7599f425e4b5a22a96bbbc509acff"} Dec 03 06:49:33 crc kubenswrapper[4818]: I1203 06:49:33.146319 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5a5a4f96-9611-456a-a084-88ba9f3b548a","Type":"ContainerStarted","Data":"523ee531ec96217d345479fc4a89b7a500f790cc222fbc60195b8b7e6ba8304a"} Dec 03 06:49:33 crc kubenswrapper[4818]: I1203 06:49:33.411317 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="8aa8deb3-c59d-4150-beb4-e68529020f60" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:36714->10.217.0.191:8775: read: connection reset by peer" Dec 03 06:49:33 crc kubenswrapper[4818]: I1203 06:49:33.411324 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="8aa8deb3-c59d-4150-beb4-e68529020f60" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:36712->10.217.0.191:8775: read: connection reset by peer" Dec 03 06:49:33 crc kubenswrapper[4818]: I1203 06:49:33.795328 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 06:49:33 crc kubenswrapper[4818]: I1203 06:49:33.965165 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-combined-ca-bundle\") pod \"8aa8deb3-c59d-4150-beb4-e68529020f60\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " Dec 03 06:49:33 crc kubenswrapper[4818]: I1203 06:49:33.965222 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-config-data\") pod \"8aa8deb3-c59d-4150-beb4-e68529020f60\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " Dec 03 06:49:33 crc kubenswrapper[4818]: I1203 06:49:33.965270 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-nova-metadata-tls-certs\") pod \"8aa8deb3-c59d-4150-beb4-e68529020f60\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " Dec 03 06:49:33 crc kubenswrapper[4818]: I1203 06:49:33.965296 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8aa8deb3-c59d-4150-beb4-e68529020f60-logs\") pod \"8aa8deb3-c59d-4150-beb4-e68529020f60\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " Dec 03 06:49:33 crc kubenswrapper[4818]: I1203 06:49:33.965316 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbmcm\" (UniqueName: \"kubernetes.io/projected/8aa8deb3-c59d-4150-beb4-e68529020f60-kube-api-access-vbmcm\") pod \"8aa8deb3-c59d-4150-beb4-e68529020f60\" (UID: \"8aa8deb3-c59d-4150-beb4-e68529020f60\") " Dec 03 06:49:33 crc kubenswrapper[4818]: I1203 06:49:33.969230 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8aa8deb3-c59d-4150-beb4-e68529020f60-logs" (OuterVolumeSpecName: "logs") pod "8aa8deb3-c59d-4150-beb4-e68529020f60" (UID: "8aa8deb3-c59d-4150-beb4-e68529020f60"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:49:33 crc kubenswrapper[4818]: I1203 06:49:33.977358 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aa8deb3-c59d-4150-beb4-e68529020f60-kube-api-access-vbmcm" (OuterVolumeSpecName: "kube-api-access-vbmcm") pod "8aa8deb3-c59d-4150-beb4-e68529020f60" (UID: "8aa8deb3-c59d-4150-beb4-e68529020f60"). InnerVolumeSpecName "kube-api-access-vbmcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:49:33 crc kubenswrapper[4818]: I1203 06:49:33.996671 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8aa8deb3-c59d-4150-beb4-e68529020f60" (UID: "8aa8deb3-c59d-4150-beb4-e68529020f60"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:33 crc kubenswrapper[4818]: I1203 06:49:33.997029 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-config-data" (OuterVolumeSpecName: "config-data") pod "8aa8deb3-c59d-4150-beb4-e68529020f60" (UID: "8aa8deb3-c59d-4150-beb4-e68529020f60"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.048237 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "8aa8deb3-c59d-4150-beb4-e68529020f60" (UID: "8aa8deb3-c59d-4150-beb4-e68529020f60"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.067912 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.067957 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.067967 4818 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8aa8deb3-c59d-4150-beb4-e68529020f60-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.067978 4818 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8aa8deb3-c59d-4150-beb4-e68529020f60-logs\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.067986 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbmcm\" (UniqueName: \"kubernetes.io/projected/8aa8deb3-c59d-4150-beb4-e68529020f60-kube-api-access-vbmcm\") on node \"crc\" DevicePath \"\"" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.177144 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"39a9a10c-e02f-4468-95a9-651b694b1f9a","Type":"ContainerStarted","Data":"b29bc003f6b864bd6d56b5559efd9c8955a7c7843317e55ed6791e259b8650f7"} Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.184581 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5a5a4f96-9611-456a-a084-88ba9f3b548a","Type":"ContainerStarted","Data":"3b1ea0c017e24ffad3ccd46626efbe23f562aba30d3d812eb3e99b41fd3f1b8d"} Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.184625 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5a5a4f96-9611-456a-a084-88ba9f3b548a","Type":"ContainerStarted","Data":"c57a25f173953b72a7c811a7ec3601fea87b23597c112b6e8fc5ff8910915773"} Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.204421 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.204397987 podStartE2EDuration="2.204397987s" podCreationTimestamp="2025-12-03 06:49:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:49:34.189234622 +0000 UTC m=+1331.880843384" watchObservedRunningTime="2025-12-03 06:49:34.204397987 +0000 UTC m=+1331.896006749" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.204914 4818 generic.go:334] "Generic (PLEG): container finished" podID="8aa8deb3-c59d-4150-beb4-e68529020f60" containerID="beab1d8ca08727518b06bd399677f6f228fac5d0e7db7c6f66ec3fedc32caf63" exitCode=0 Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.204961 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8aa8deb3-c59d-4150-beb4-e68529020f60","Type":"ContainerDied","Data":"beab1d8ca08727518b06bd399677f6f228fac5d0e7db7c6f66ec3fedc32caf63"} Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.204993 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8aa8deb3-c59d-4150-beb4-e68529020f60","Type":"ContainerDied","Data":"ab8337e4b0156c9798308173a72daf3ad0da1540018a259c392605601a958682"} Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.205015 4818 scope.go:117] "RemoveContainer" containerID="beab1d8ca08727518b06bd399677f6f228fac5d0e7db7c6f66ec3fedc32caf63" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.205098 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.220642 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.220620238 podStartE2EDuration="2.220620238s" podCreationTimestamp="2025-12-03 06:49:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:49:34.209347449 +0000 UTC m=+1331.900956201" watchObservedRunningTime="2025-12-03 06:49:34.220620238 +0000 UTC m=+1331.912228990" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.247604 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.251493 4818 scope.go:117] "RemoveContainer" containerID="af1f3345b5796f8d0b20c74bc648f214297ce4edb68bdd11ad027e991ae2237a" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.275465 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.289942 4818 scope.go:117] "RemoveContainer" containerID="beab1d8ca08727518b06bd399677f6f228fac5d0e7db7c6f66ec3fedc32caf63" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.291020 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:49:34 crc kubenswrapper[4818]: E1203 06:49:34.291486 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aa8deb3-c59d-4150-beb4-e68529020f60" containerName="nova-metadata-log" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.291512 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aa8deb3-c59d-4150-beb4-e68529020f60" containerName="nova-metadata-log" Dec 03 06:49:34 crc kubenswrapper[4818]: E1203 06:49:34.291537 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aa8deb3-c59d-4150-beb4-e68529020f60" containerName="nova-metadata-metadata" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.291547 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aa8deb3-c59d-4150-beb4-e68529020f60" containerName="nova-metadata-metadata" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.291789 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aa8deb3-c59d-4150-beb4-e68529020f60" containerName="nova-metadata-metadata" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.291843 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aa8deb3-c59d-4150-beb4-e68529020f60" containerName="nova-metadata-log" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.293118 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: E1203 06:49:34.295143 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"beab1d8ca08727518b06bd399677f6f228fac5d0e7db7c6f66ec3fedc32caf63\": container with ID starting with beab1d8ca08727518b06bd399677f6f228fac5d0e7db7c6f66ec3fedc32caf63 not found: ID does not exist" containerID="beab1d8ca08727518b06bd399677f6f228fac5d0e7db7c6f66ec3fedc32caf63" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.295183 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beab1d8ca08727518b06bd399677f6f228fac5d0e7db7c6f66ec3fedc32caf63"} err="failed to get container status \"beab1d8ca08727518b06bd399677f6f228fac5d0e7db7c6f66ec3fedc32caf63\": rpc error: code = NotFound desc = could not find container \"beab1d8ca08727518b06bd399677f6f228fac5d0e7db7c6f66ec3fedc32caf63\": container with ID starting with beab1d8ca08727518b06bd399677f6f228fac5d0e7db7c6f66ec3fedc32caf63 not found: ID does not exist" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.295214 4818 scope.go:117] "RemoveContainer" containerID="af1f3345b5796f8d0b20c74bc648f214297ce4edb68bdd11ad027e991ae2237a" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.297691 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.297798 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 06:49:34 crc kubenswrapper[4818]: E1203 06:49:34.299465 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af1f3345b5796f8d0b20c74bc648f214297ce4edb68bdd11ad027e991ae2237a\": container with ID starting with af1f3345b5796f8d0b20c74bc648f214297ce4edb68bdd11ad027e991ae2237a not found: ID does not exist" containerID="af1f3345b5796f8d0b20c74bc648f214297ce4edb68bdd11ad027e991ae2237a" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.299508 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af1f3345b5796f8d0b20c74bc648f214297ce4edb68bdd11ad027e991ae2237a"} err="failed to get container status \"af1f3345b5796f8d0b20c74bc648f214297ce4edb68bdd11ad027e991ae2237a\": rpc error: code = NotFound desc = could not find container \"af1f3345b5796f8d0b20c74bc648f214297ce4edb68bdd11ad027e991ae2237a\": container with ID starting with af1f3345b5796f8d0b20c74bc648f214297ce4edb68bdd11ad027e991ae2237a not found: ID does not exist" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.302439 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.378530 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78622406-d448-4cf4-b5c9-c639b0fb6a41-config-data\") pod \"nova-metadata-0\" (UID: \"78622406-d448-4cf4-b5c9-c639b0fb6a41\") " pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.378583 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/78622406-d448-4cf4-b5c9-c639b0fb6a41-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"78622406-d448-4cf4-b5c9-c639b0fb6a41\") " pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.378615 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78622406-d448-4cf4-b5c9-c639b0fb6a41-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"78622406-d448-4cf4-b5c9-c639b0fb6a41\") " pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.378704 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78622406-d448-4cf4-b5c9-c639b0fb6a41-logs\") pod \"nova-metadata-0\" (UID: \"78622406-d448-4cf4-b5c9-c639b0fb6a41\") " pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.378826 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqf4d\" (UniqueName: \"kubernetes.io/projected/78622406-d448-4cf4-b5c9-c639b0fb6a41-kube-api-access-hqf4d\") pod \"nova-metadata-0\" (UID: \"78622406-d448-4cf4-b5c9-c639b0fb6a41\") " pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.480314 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78622406-d448-4cf4-b5c9-c639b0fb6a41-config-data\") pod \"nova-metadata-0\" (UID: \"78622406-d448-4cf4-b5c9-c639b0fb6a41\") " pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.480384 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/78622406-d448-4cf4-b5c9-c639b0fb6a41-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"78622406-d448-4cf4-b5c9-c639b0fb6a41\") " pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.480420 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78622406-d448-4cf4-b5c9-c639b0fb6a41-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"78622406-d448-4cf4-b5c9-c639b0fb6a41\") " pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.480465 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78622406-d448-4cf4-b5c9-c639b0fb6a41-logs\") pod \"nova-metadata-0\" (UID: \"78622406-d448-4cf4-b5c9-c639b0fb6a41\") " pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.480578 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqf4d\" (UniqueName: \"kubernetes.io/projected/78622406-d448-4cf4-b5c9-c639b0fb6a41-kube-api-access-hqf4d\") pod \"nova-metadata-0\" (UID: \"78622406-d448-4cf4-b5c9-c639b0fb6a41\") " pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.481126 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78622406-d448-4cf4-b5c9-c639b0fb6a41-logs\") pod \"nova-metadata-0\" (UID: \"78622406-d448-4cf4-b5c9-c639b0fb6a41\") " pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.484062 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78622406-d448-4cf4-b5c9-c639b0fb6a41-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"78622406-d448-4cf4-b5c9-c639b0fb6a41\") " pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.490481 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/78622406-d448-4cf4-b5c9-c639b0fb6a41-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"78622406-d448-4cf4-b5c9-c639b0fb6a41\") " pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.490596 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78622406-d448-4cf4-b5c9-c639b0fb6a41-config-data\") pod \"nova-metadata-0\" (UID: \"78622406-d448-4cf4-b5c9-c639b0fb6a41\") " pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.498321 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqf4d\" (UniqueName: \"kubernetes.io/projected/78622406-d448-4cf4-b5c9-c639b0fb6a41-kube-api-access-hqf4d\") pod \"nova-metadata-0\" (UID: \"78622406-d448-4cf4-b5c9-c639b0fb6a41\") " pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.612015 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 06:49:34 crc kubenswrapper[4818]: I1203 06:49:34.759640 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8aa8deb3-c59d-4150-beb4-e68529020f60" path="/var/lib/kubelet/pods/8aa8deb3-c59d-4150-beb4-e68529020f60/volumes" Dec 03 06:49:35 crc kubenswrapper[4818]: I1203 06:49:35.097718 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 06:49:35 crc kubenswrapper[4818]: W1203 06:49:35.099907 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78622406_d448_4cf4_b5c9_c639b0fb6a41.slice/crio-6d9f9b082f740f15f47dd0d57cd81b09424a1b398f20b4cc1988bce14f06543f WatchSource:0}: Error finding container 6d9f9b082f740f15f47dd0d57cd81b09424a1b398f20b4cc1988bce14f06543f: Status 404 returned error can't find the container with id 6d9f9b082f740f15f47dd0d57cd81b09424a1b398f20b4cc1988bce14f06543f Dec 03 06:49:35 crc kubenswrapper[4818]: I1203 06:49:35.216511 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"78622406-d448-4cf4-b5c9-c639b0fb6a41","Type":"ContainerStarted","Data":"6d9f9b082f740f15f47dd0d57cd81b09424a1b398f20b4cc1988bce14f06543f"} Dec 03 06:49:36 crc kubenswrapper[4818]: I1203 06:49:36.248603 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"78622406-d448-4cf4-b5c9-c639b0fb6a41","Type":"ContainerStarted","Data":"0d8bf2a5a04359a66d40991f6bd0e753b71ef9582cc4d74b2b337e39c1c54437"} Dec 03 06:49:36 crc kubenswrapper[4818]: I1203 06:49:36.249394 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"78622406-d448-4cf4-b5c9-c639b0fb6a41","Type":"ContainerStarted","Data":"fffe4a6cebdfd1ececaefe05a048ae5f76b4b2c63401ee0b569f73010cff3e5b"} Dec 03 06:49:36 crc kubenswrapper[4818]: I1203 06:49:36.293543 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.293517537 podStartE2EDuration="2.293517537s" podCreationTimestamp="2025-12-03 06:49:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:49:36.277807099 +0000 UTC m=+1333.969415871" watchObservedRunningTime="2025-12-03 06:49:36.293517537 +0000 UTC m=+1333.985126299" Dec 03 06:49:37 crc kubenswrapper[4818]: I1203 06:49:37.590053 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 06:49:39 crc kubenswrapper[4818]: I1203 06:49:39.612368 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 06:49:39 crc kubenswrapper[4818]: I1203 06:49:39.612674 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 06:49:42 crc kubenswrapper[4818]: I1203 06:49:42.581792 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 06:49:42 crc kubenswrapper[4818]: I1203 06:49:42.582290 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 06:49:42 crc kubenswrapper[4818]: I1203 06:49:42.591034 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 06:49:42 crc kubenswrapper[4818]: I1203 06:49:42.617295 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 06:49:43 crc kubenswrapper[4818]: I1203 06:49:43.302270 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:49:43 crc kubenswrapper[4818]: I1203 06:49:43.302331 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:49:43 crc kubenswrapper[4818]: I1203 06:49:43.418627 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 06:49:43 crc kubenswrapper[4818]: I1203 06:49:43.626977 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5a5a4f96-9611-456a-a084-88ba9f3b548a" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 06:49:43 crc kubenswrapper[4818]: I1203 06:49:43.627015 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5a5a4f96-9611-456a-a084-88ba9f3b548a" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 06:49:44 crc kubenswrapper[4818]: I1203 06:49:44.612895 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 06:49:44 crc kubenswrapper[4818]: I1203 06:49:44.613441 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 06:49:45 crc kubenswrapper[4818]: I1203 06:49:45.650052 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="78622406-d448-4cf4-b5c9-c639b0fb6a41" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 06:49:45 crc kubenswrapper[4818]: I1203 06:49:45.661204 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="78622406-d448-4cf4-b5c9-c639b0fb6a41" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 06:49:49 crc kubenswrapper[4818]: I1203 06:49:49.379217 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 06:49:52 crc kubenswrapper[4818]: I1203 06:49:52.590155 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 06:49:52 crc kubenswrapper[4818]: I1203 06:49:52.592938 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 06:49:52 crc kubenswrapper[4818]: I1203 06:49:52.600769 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 06:49:52 crc kubenswrapper[4818]: I1203 06:49:52.785750 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 06:49:53 crc kubenswrapper[4818]: I1203 06:49:53.502978 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 06:49:53 crc kubenswrapper[4818]: I1203 06:49:53.510957 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 06:49:54 crc kubenswrapper[4818]: I1203 06:49:54.618189 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 06:49:54 crc kubenswrapper[4818]: I1203 06:49:54.621350 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 06:49:54 crc kubenswrapper[4818]: I1203 06:49:54.622877 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 06:49:55 crc kubenswrapper[4818]: I1203 06:49:55.526920 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 06:50:04 crc kubenswrapper[4818]: I1203 06:50:04.314185 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 06:50:05 crc kubenswrapper[4818]: I1203 06:50:05.902582 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 06:50:08 crc kubenswrapper[4818]: I1203 06:50:08.863995 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="0e751691-c660-4527-a9a6-61505f83c6ff" containerName="rabbitmq" containerID="cri-o://5b437a91614dddcbffaba1e3457305c6be3910b1ccccef8d5c8a44875396d9f1" gracePeriod=604796 Dec 03 06:50:10 crc kubenswrapper[4818]: I1203 06:50:10.144134 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="df9ab7c7-152b-4938-9ef5-d0588111b8a2" containerName="rabbitmq" containerID="cri-o://b14a6d3062e8c0e207497d3cecb33774c99ddc3ee3d60372949002a157583eed" gracePeriod=604796 Dec 03 06:50:13 crc kubenswrapper[4818]: I1203 06:50:13.302458 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:50:13 crc kubenswrapper[4818]: I1203 06:50:13.303866 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:50:13 crc kubenswrapper[4818]: I1203 06:50:13.304061 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:50:13 crc kubenswrapper[4818]: I1203 06:50:13.305081 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5bf3e0e65704b43280754ba18a3250980330aab9549a78ae7d9381e3d0130a33"} pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 06:50:13 crc kubenswrapper[4818]: I1203 06:50:13.305301 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" containerID="cri-o://5bf3e0e65704b43280754ba18a3250980330aab9549a78ae7d9381e3d0130a33" gracePeriod=600 Dec 03 06:50:13 crc kubenswrapper[4818]: I1203 06:50:13.756267 4818 generic.go:334] "Generic (PLEG): container finished" podID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerID="5bf3e0e65704b43280754ba18a3250980330aab9549a78ae7d9381e3d0130a33" exitCode=0 Dec 03 06:50:13 crc kubenswrapper[4818]: I1203 06:50:13.756337 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerDied","Data":"5bf3e0e65704b43280754ba18a3250980330aab9549a78ae7d9381e3d0130a33"} Dec 03 06:50:13 crc kubenswrapper[4818]: I1203 06:50:13.756668 4818 scope.go:117] "RemoveContainer" containerID="94f995293ac99f0a139bded43f6e45522bd84517d494216318f538517e1db600" Dec 03 06:50:14 crc kubenswrapper[4818]: I1203 06:50:14.768332 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01"} Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.445991 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.554605 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-confd\") pod \"0e751691-c660-4527-a9a6-61505f83c6ff\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.554687 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-config-data\") pod \"0e751691-c660-4527-a9a6-61505f83c6ff\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.554737 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-plugins-conf\") pod \"0e751691-c660-4527-a9a6-61505f83c6ff\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.554903 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-erlang-cookie\") pod \"0e751691-c660-4527-a9a6-61505f83c6ff\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.555029 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbbg7\" (UniqueName: \"kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-kube-api-access-dbbg7\") pod \"0e751691-c660-4527-a9a6-61505f83c6ff\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.555109 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-plugins\") pod \"0e751691-c660-4527-a9a6-61505f83c6ff\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.555203 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0e751691-c660-4527-a9a6-61505f83c6ff-erlang-cookie-secret\") pod \"0e751691-c660-4527-a9a6-61505f83c6ff\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.555239 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"0e751691-c660-4527-a9a6-61505f83c6ff\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.555295 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0e751691-c660-4527-a9a6-61505f83c6ff-pod-info\") pod \"0e751691-c660-4527-a9a6-61505f83c6ff\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.555352 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-server-conf\") pod \"0e751691-c660-4527-a9a6-61505f83c6ff\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.555400 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-tls\") pod \"0e751691-c660-4527-a9a6-61505f83c6ff\" (UID: \"0e751691-c660-4527-a9a6-61505f83c6ff\") " Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.558168 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0e751691-c660-4527-a9a6-61505f83c6ff" (UID: "0e751691-c660-4527-a9a6-61505f83c6ff"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.559583 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0e751691-c660-4527-a9a6-61505f83c6ff" (UID: "0e751691-c660-4527-a9a6-61505f83c6ff"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.560531 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0e751691-c660-4527-a9a6-61505f83c6ff" (UID: "0e751691-c660-4527-a9a6-61505f83c6ff"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.568928 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e751691-c660-4527-a9a6-61505f83c6ff-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0e751691-c660-4527-a9a6-61505f83c6ff" (UID: "0e751691-c660-4527-a9a6-61505f83c6ff"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.576470 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0e751691-c660-4527-a9a6-61505f83c6ff" (UID: "0e751691-c660-4527-a9a6-61505f83c6ff"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.576671 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "0e751691-c660-4527-a9a6-61505f83c6ff" (UID: "0e751691-c660-4527-a9a6-61505f83c6ff"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.577145 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-kube-api-access-dbbg7" (OuterVolumeSpecName: "kube-api-access-dbbg7") pod "0e751691-c660-4527-a9a6-61505f83c6ff" (UID: "0e751691-c660-4527-a9a6-61505f83c6ff"). InnerVolumeSpecName "kube-api-access-dbbg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.580456 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0e751691-c660-4527-a9a6-61505f83c6ff-pod-info" (OuterVolumeSpecName: "pod-info") pod "0e751691-c660-4527-a9a6-61505f83c6ff" (UID: "0e751691-c660-4527-a9a6-61505f83c6ff"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.599588 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-config-data" (OuterVolumeSpecName: "config-data") pod "0e751691-c660-4527-a9a6-61505f83c6ff" (UID: "0e751691-c660-4527-a9a6-61505f83c6ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.633179 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-server-conf" (OuterVolumeSpecName: "server-conf") pod "0e751691-c660-4527-a9a6-61505f83c6ff" (UID: "0e751691-c660-4527-a9a6-61505f83c6ff"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.657483 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.657524 4818 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.657539 4818 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.657551 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbbg7\" (UniqueName: \"kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-kube-api-access-dbbg7\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.657561 4818 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.657569 4818 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0e751691-c660-4527-a9a6-61505f83c6ff-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.657592 4818 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.657601 4818 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0e751691-c660-4527-a9a6-61505f83c6ff-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.657608 4818 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0e751691-c660-4527-a9a6-61505f83c6ff-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.657617 4818 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.687543 4818 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.741769 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0e751691-c660-4527-a9a6-61505f83c6ff" (UID: "0e751691-c660-4527-a9a6-61505f83c6ff"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.760126 4818 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.761745 4818 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0e751691-c660-4527-a9a6-61505f83c6ff-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.778187 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.778190 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0e751691-c660-4527-a9a6-61505f83c6ff","Type":"ContainerDied","Data":"5b437a91614dddcbffaba1e3457305c6be3910b1ccccef8d5c8a44875396d9f1"} Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.778312 4818 scope.go:117] "RemoveContainer" containerID="5b437a91614dddcbffaba1e3457305c6be3910b1ccccef8d5c8a44875396d9f1" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.779060 4818 generic.go:334] "Generic (PLEG): container finished" podID="0e751691-c660-4527-a9a6-61505f83c6ff" containerID="5b437a91614dddcbffaba1e3457305c6be3910b1ccccef8d5c8a44875396d9f1" exitCode=0 Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.779136 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0e751691-c660-4527-a9a6-61505f83c6ff","Type":"ContainerDied","Data":"7ee3814f8164ad501c7750b02416b4c23efac5aded2434000a581b60871030d6"} Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.804969 4818 scope.go:117] "RemoveContainer" containerID="303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.821049 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.827893 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.847693 4818 scope.go:117] "RemoveContainer" containerID="5b437a91614dddcbffaba1e3457305c6be3910b1ccccef8d5c8a44875396d9f1" Dec 03 06:50:15 crc kubenswrapper[4818]: E1203 06:50:15.848284 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b437a91614dddcbffaba1e3457305c6be3910b1ccccef8d5c8a44875396d9f1\": container with ID starting with 5b437a91614dddcbffaba1e3457305c6be3910b1ccccef8d5c8a44875396d9f1 not found: ID does not exist" containerID="5b437a91614dddcbffaba1e3457305c6be3910b1ccccef8d5c8a44875396d9f1" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.848325 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b437a91614dddcbffaba1e3457305c6be3910b1ccccef8d5c8a44875396d9f1"} err="failed to get container status \"5b437a91614dddcbffaba1e3457305c6be3910b1ccccef8d5c8a44875396d9f1\": rpc error: code = NotFound desc = could not find container \"5b437a91614dddcbffaba1e3457305c6be3910b1ccccef8d5c8a44875396d9f1\": container with ID starting with 5b437a91614dddcbffaba1e3457305c6be3910b1ccccef8d5c8a44875396d9f1 not found: ID does not exist" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.848350 4818 scope.go:117] "RemoveContainer" containerID="303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0" Dec 03 06:50:15 crc kubenswrapper[4818]: E1203 06:50:15.848751 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0\": container with ID starting with 303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0 not found: ID does not exist" containerID="303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.848778 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0"} err="failed to get container status \"303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0\": rpc error: code = NotFound desc = could not find container \"303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0\": container with ID starting with 303718f5f9011eaa81656dfc886069404b711f73c70d38d97fe3fb9b491d09e0 not found: ID does not exist" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.853841 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 06:50:15 crc kubenswrapper[4818]: E1203 06:50:15.854322 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e751691-c660-4527-a9a6-61505f83c6ff" containerName="rabbitmq" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.854349 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e751691-c660-4527-a9a6-61505f83c6ff" containerName="rabbitmq" Dec 03 06:50:15 crc kubenswrapper[4818]: E1203 06:50:15.854360 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e751691-c660-4527-a9a6-61505f83c6ff" containerName="setup-container" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.854368 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e751691-c660-4527-a9a6-61505f83c6ff" containerName="setup-container" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.854623 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e751691-c660-4527-a9a6-61505f83c6ff" containerName="rabbitmq" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.856020 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.858263 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.858553 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.858753 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.858796 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.859110 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.859146 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.859267 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-jws4q" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.874099 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.964903 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e6fa481e-32ce-4840-b457-0a692558cdb5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.964985 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e6fa481e-32ce-4840-b457-0a692558cdb5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.965027 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e6fa481e-32ce-4840-b457-0a692558cdb5-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.965054 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e6fa481e-32ce-4840-b457-0a692558cdb5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.965074 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e6fa481e-32ce-4840-b457-0a692558cdb5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.965115 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e6fa481e-32ce-4840-b457-0a692558cdb5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.965134 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e6fa481e-32ce-4840-b457-0a692558cdb5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.965152 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e6fa481e-32ce-4840-b457-0a692558cdb5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.965187 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.965217 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncm9x\" (UniqueName: \"kubernetes.io/projected/e6fa481e-32ce-4840-b457-0a692558cdb5-kube-api-access-ncm9x\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:15 crc kubenswrapper[4818]: I1203 06:50:15.965239 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e6fa481e-32ce-4840-b457-0a692558cdb5-config-data\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.066801 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e6fa481e-32ce-4840-b457-0a692558cdb5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.066894 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e6fa481e-32ce-4840-b457-0a692558cdb5-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.066928 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e6fa481e-32ce-4840-b457-0a692558cdb5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.066985 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e6fa481e-32ce-4840-b457-0a692558cdb5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.067032 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e6fa481e-32ce-4840-b457-0a692558cdb5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.067050 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e6fa481e-32ce-4840-b457-0a692558cdb5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.067072 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e6fa481e-32ce-4840-b457-0a692558cdb5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.067109 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.067150 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncm9x\" (UniqueName: \"kubernetes.io/projected/e6fa481e-32ce-4840-b457-0a692558cdb5-kube-api-access-ncm9x\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.067175 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e6fa481e-32ce-4840-b457-0a692558cdb5-config-data\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.067200 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e6fa481e-32ce-4840-b457-0a692558cdb5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.067684 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.067718 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e6fa481e-32ce-4840-b457-0a692558cdb5-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.068336 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e6fa481e-32ce-4840-b457-0a692558cdb5-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.068395 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e6fa481e-32ce-4840-b457-0a692558cdb5-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.068852 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e6fa481e-32ce-4840-b457-0a692558cdb5-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.069015 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e6fa481e-32ce-4840-b457-0a692558cdb5-config-data\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.071250 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e6fa481e-32ce-4840-b457-0a692558cdb5-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.071345 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e6fa481e-32ce-4840-b457-0a692558cdb5-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.071507 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e6fa481e-32ce-4840-b457-0a692558cdb5-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.071886 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e6fa481e-32ce-4840-b457-0a692558cdb5-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.088753 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncm9x\" (UniqueName: \"kubernetes.io/projected/e6fa481e-32ce-4840-b457-0a692558cdb5-kube-api-access-ncm9x\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.101639 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"e6fa481e-32ce-4840-b457-0a692558cdb5\") " pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.179566 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.256741 4818 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="df9ab7c7-152b-4938-9ef5-d0588111b8a2" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.696227 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.711633 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 06:50:16 crc kubenswrapper[4818]: W1203 06:50:16.712882 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6fa481e_32ce_4840_b457_0a692558cdb5.slice/crio-25d58eccbe57d31c6250d31fb5c641653c065505b024939212be3562b0126ffa WatchSource:0}: Error finding container 25d58eccbe57d31c6250d31fb5c641653c065505b024939212be3562b0126ffa: Status 404 returned error can't find the container with id 25d58eccbe57d31c6250d31fb5c641653c065505b024939212be3562b0126ffa Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.779245 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e751691-c660-4527-a9a6-61505f83c6ff" path="/var/lib/kubelet/pods/0e751691-c660-4527-a9a6-61505f83c6ff/volumes" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.811717 4818 generic.go:334] "Generic (PLEG): container finished" podID="df9ab7c7-152b-4938-9ef5-d0588111b8a2" containerID="b14a6d3062e8c0e207497d3cecb33774c99ddc3ee3d60372949002a157583eed" exitCode=0 Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.812198 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.812258 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"df9ab7c7-152b-4938-9ef5-d0588111b8a2","Type":"ContainerDied","Data":"b14a6d3062e8c0e207497d3cecb33774c99ddc3ee3d60372949002a157583eed"} Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.812303 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"df9ab7c7-152b-4938-9ef5-d0588111b8a2","Type":"ContainerDied","Data":"13eb95568ccbe82db9fe67086da897b8d23545d092808a160162c9904584eeda"} Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.812321 4818 scope.go:117] "RemoveContainer" containerID="b14a6d3062e8c0e207497d3cecb33774c99ddc3ee3d60372949002a157583eed" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.827098 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e6fa481e-32ce-4840-b457-0a692558cdb5","Type":"ContainerStarted","Data":"25d58eccbe57d31c6250d31fb5c641653c065505b024939212be3562b0126ffa"} Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.854384 4818 scope.go:117] "RemoveContainer" containerID="651b7927b0f2dd739e06ae0b1b47c6bf30fb2fa4bce7e634b77e016a5d7e4fce" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.877746 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pwn2\" (UniqueName: \"kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-kube-api-access-8pwn2\") pod \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.877851 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-erlang-cookie\") pod \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.877883 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-confd\") pod \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.877910 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df9ab7c7-152b-4938-9ef5-d0588111b8a2-erlang-cookie-secret\") pod \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.877941 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df9ab7c7-152b-4938-9ef5-d0588111b8a2-pod-info\") pod \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.877971 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-server-conf\") pod \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.878001 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-plugins\") pod \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.878024 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-plugins-conf\") pod \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.878075 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-tls\") pod \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.878164 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.878189 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-config-data\") pod \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\" (UID: \"df9ab7c7-152b-4938-9ef5-d0588111b8a2\") " Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.878731 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "df9ab7c7-152b-4938-9ef5-d0588111b8a2" (UID: "df9ab7c7-152b-4938-9ef5-d0588111b8a2"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.879961 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "df9ab7c7-152b-4938-9ef5-d0588111b8a2" (UID: "df9ab7c7-152b-4938-9ef5-d0588111b8a2"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.880077 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "df9ab7c7-152b-4938-9ef5-d0588111b8a2" (UID: "df9ab7c7-152b-4938-9ef5-d0588111b8a2"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.883449 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "df9ab7c7-152b-4938-9ef5-d0588111b8a2" (UID: "df9ab7c7-152b-4938-9ef5-d0588111b8a2"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.885466 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/df9ab7c7-152b-4938-9ef5-d0588111b8a2-pod-info" (OuterVolumeSpecName: "pod-info") pod "df9ab7c7-152b-4938-9ef5-d0588111b8a2" (UID: "df9ab7c7-152b-4938-9ef5-d0588111b8a2"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.885626 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-kube-api-access-8pwn2" (OuterVolumeSpecName: "kube-api-access-8pwn2") pod "df9ab7c7-152b-4938-9ef5-d0588111b8a2" (UID: "df9ab7c7-152b-4938-9ef5-d0588111b8a2"). InnerVolumeSpecName "kube-api-access-8pwn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.885651 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df9ab7c7-152b-4938-9ef5-d0588111b8a2-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "df9ab7c7-152b-4938-9ef5-d0588111b8a2" (UID: "df9ab7c7-152b-4938-9ef5-d0588111b8a2"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.886762 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "df9ab7c7-152b-4938-9ef5-d0588111b8a2" (UID: "df9ab7c7-152b-4938-9ef5-d0588111b8a2"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.890581 4818 scope.go:117] "RemoveContainer" containerID="b14a6d3062e8c0e207497d3cecb33774c99ddc3ee3d60372949002a157583eed" Dec 03 06:50:16 crc kubenswrapper[4818]: E1203 06:50:16.891001 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b14a6d3062e8c0e207497d3cecb33774c99ddc3ee3d60372949002a157583eed\": container with ID starting with b14a6d3062e8c0e207497d3cecb33774c99ddc3ee3d60372949002a157583eed not found: ID does not exist" containerID="b14a6d3062e8c0e207497d3cecb33774c99ddc3ee3d60372949002a157583eed" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.891033 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b14a6d3062e8c0e207497d3cecb33774c99ddc3ee3d60372949002a157583eed"} err="failed to get container status \"b14a6d3062e8c0e207497d3cecb33774c99ddc3ee3d60372949002a157583eed\": rpc error: code = NotFound desc = could not find container \"b14a6d3062e8c0e207497d3cecb33774c99ddc3ee3d60372949002a157583eed\": container with ID starting with b14a6d3062e8c0e207497d3cecb33774c99ddc3ee3d60372949002a157583eed not found: ID does not exist" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.891060 4818 scope.go:117] "RemoveContainer" containerID="651b7927b0f2dd739e06ae0b1b47c6bf30fb2fa4bce7e634b77e016a5d7e4fce" Dec 03 06:50:16 crc kubenswrapper[4818]: E1203 06:50:16.891313 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"651b7927b0f2dd739e06ae0b1b47c6bf30fb2fa4bce7e634b77e016a5d7e4fce\": container with ID starting with 651b7927b0f2dd739e06ae0b1b47c6bf30fb2fa4bce7e634b77e016a5d7e4fce not found: ID does not exist" containerID="651b7927b0f2dd739e06ae0b1b47c6bf30fb2fa4bce7e634b77e016a5d7e4fce" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.891338 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"651b7927b0f2dd739e06ae0b1b47c6bf30fb2fa4bce7e634b77e016a5d7e4fce"} err="failed to get container status \"651b7927b0f2dd739e06ae0b1b47c6bf30fb2fa4bce7e634b77e016a5d7e4fce\": rpc error: code = NotFound desc = could not find container \"651b7927b0f2dd739e06ae0b1b47c6bf30fb2fa4bce7e634b77e016a5d7e4fce\": container with ID starting with 651b7927b0f2dd739e06ae0b1b47c6bf30fb2fa4bce7e634b77e016a5d7e4fce not found: ID does not exist" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.915019 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-config-data" (OuterVolumeSpecName: "config-data") pod "df9ab7c7-152b-4938-9ef5-d0588111b8a2" (UID: "df9ab7c7-152b-4938-9ef5-d0588111b8a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.943894 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-server-conf" (OuterVolumeSpecName: "server-conf") pod "df9ab7c7-152b-4938-9ef5-d0588111b8a2" (UID: "df9ab7c7-152b-4938-9ef5-d0588111b8a2"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.980714 4818 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.980742 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.980753 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pwn2\" (UniqueName: \"kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-kube-api-access-8pwn2\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.980764 4818 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.980772 4818 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df9ab7c7-152b-4938-9ef5-d0588111b8a2-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.980780 4818 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df9ab7c7-152b-4938-9ef5-d0588111b8a2-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.980788 4818 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.980796 4818 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.980804 4818 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df9ab7c7-152b-4938-9ef5-d0588111b8a2-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.980825 4818 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.981023 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "df9ab7c7-152b-4938-9ef5-d0588111b8a2" (UID: "df9ab7c7-152b-4938-9ef5-d0588111b8a2"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:16 crc kubenswrapper[4818]: I1203 06:50:16.998651 4818 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.082976 4818 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df9ab7c7-152b-4938-9ef5-d0588111b8a2-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.083022 4818 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.173987 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.186850 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.198506 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 06:50:17 crc kubenswrapper[4818]: E1203 06:50:17.199020 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df9ab7c7-152b-4938-9ef5-d0588111b8a2" containerName="setup-container" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.199042 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="df9ab7c7-152b-4938-9ef5-d0588111b8a2" containerName="setup-container" Dec 03 06:50:17 crc kubenswrapper[4818]: E1203 06:50:17.199056 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df9ab7c7-152b-4938-9ef5-d0588111b8a2" containerName="rabbitmq" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.199064 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="df9ab7c7-152b-4938-9ef5-d0588111b8a2" containerName="rabbitmq" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.199300 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="df9ab7c7-152b-4938-9ef5-d0588111b8a2" containerName="rabbitmq" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.200376 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.203559 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.203772 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-pnlrk" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.204056 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.204240 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.206465 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.206979 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.211437 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.219954 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.388227 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8078039a-8674-40c1-ba1a-ff8244d17e16-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.388657 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8078039a-8674-40c1-ba1a-ff8244d17e16-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.388843 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8078039a-8674-40c1-ba1a-ff8244d17e16-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.388955 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.389245 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8078039a-8674-40c1-ba1a-ff8244d17e16-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.389363 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8078039a-8674-40c1-ba1a-ff8244d17e16-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.389486 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8078039a-8674-40c1-ba1a-ff8244d17e16-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.389608 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj7pc\" (UniqueName: \"kubernetes.io/projected/8078039a-8674-40c1-ba1a-ff8244d17e16-kube-api-access-lj7pc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.389724 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8078039a-8674-40c1-ba1a-ff8244d17e16-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.389912 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8078039a-8674-40c1-ba1a-ff8244d17e16-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.390021 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8078039a-8674-40c1-ba1a-ff8244d17e16-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.491288 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8078039a-8674-40c1-ba1a-ff8244d17e16-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.491332 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8078039a-8674-40c1-ba1a-ff8244d17e16-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.491368 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8078039a-8674-40c1-ba1a-ff8244d17e16-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.491389 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.491441 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8078039a-8674-40c1-ba1a-ff8244d17e16-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.491467 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8078039a-8674-40c1-ba1a-ff8244d17e16-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.491494 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8078039a-8674-40c1-ba1a-ff8244d17e16-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.491527 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj7pc\" (UniqueName: \"kubernetes.io/projected/8078039a-8674-40c1-ba1a-ff8244d17e16-kube-api-access-lj7pc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.491548 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8078039a-8674-40c1-ba1a-ff8244d17e16-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.491598 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8078039a-8674-40c1-ba1a-ff8244d17e16-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.491619 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8078039a-8674-40c1-ba1a-ff8244d17e16-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.491785 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.492061 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8078039a-8674-40c1-ba1a-ff8244d17e16-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.492255 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8078039a-8674-40c1-ba1a-ff8244d17e16-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.492358 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8078039a-8674-40c1-ba1a-ff8244d17e16-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.493506 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8078039a-8674-40c1-ba1a-ff8244d17e16-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.494201 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8078039a-8674-40c1-ba1a-ff8244d17e16-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.497012 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8078039a-8674-40c1-ba1a-ff8244d17e16-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.497631 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8078039a-8674-40c1-ba1a-ff8244d17e16-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.498012 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8078039a-8674-40c1-ba1a-ff8244d17e16-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.511594 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8078039a-8674-40c1-ba1a-ff8244d17e16-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.514588 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj7pc\" (UniqueName: \"kubernetes.io/projected/8078039a-8674-40c1-ba1a-ff8244d17e16-kube-api-access-lj7pc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.527855 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"8078039a-8674-40c1-ba1a-ff8244d17e16\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:17 crc kubenswrapper[4818]: I1203 06:50:17.821169 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:18 crc kubenswrapper[4818]: I1203 06:50:18.288292 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 06:50:18 crc kubenswrapper[4818]: W1203 06:50:18.289298 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8078039a_8674_40c1_ba1a_ff8244d17e16.slice/crio-23fcce37061fd8799dce61b0fe556425918109cc8ead853ee77fa48acfcbefdd WatchSource:0}: Error finding container 23fcce37061fd8799dce61b0fe556425918109cc8ead853ee77fa48acfcbefdd: Status 404 returned error can't find the container with id 23fcce37061fd8799dce61b0fe556425918109cc8ead853ee77fa48acfcbefdd Dec 03 06:50:18 crc kubenswrapper[4818]: I1203 06:50:18.750345 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df9ab7c7-152b-4938-9ef5-d0588111b8a2" path="/var/lib/kubelet/pods/df9ab7c7-152b-4938-9ef5-d0588111b8a2/volumes" Dec 03 06:50:18 crc kubenswrapper[4818]: I1203 06:50:18.788543 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-lgdlb"] Dec 03 06:50:18 crc kubenswrapper[4818]: I1203 06:50:18.790096 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:18 crc kubenswrapper[4818]: I1203 06:50:18.801077 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-lgdlb"] Dec 03 06:50:18 crc kubenswrapper[4818]: I1203 06:50:18.857654 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 03 06:50:18 crc kubenswrapper[4818]: I1203 06:50:18.872706 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e6fa481e-32ce-4840-b457-0a692558cdb5","Type":"ContainerStarted","Data":"1bed1e3c0e6ad25d79ffed0a1e1704cf610cbc76ef15c729896c79167c93693f"} Dec 03 06:50:18 crc kubenswrapper[4818]: I1203 06:50:18.876092 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8078039a-8674-40c1-ba1a-ff8244d17e16","Type":"ContainerStarted","Data":"23fcce37061fd8799dce61b0fe556425918109cc8ead853ee77fa48acfcbefdd"} Dec 03 06:50:18 crc kubenswrapper[4818]: I1203 06:50:18.961528 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:18 crc kubenswrapper[4818]: I1203 06:50:18.961663 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk54h\" (UniqueName: \"kubernetes.io/projected/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-kube-api-access-lk54h\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:18 crc kubenswrapper[4818]: I1203 06:50:18.961691 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:18 crc kubenswrapper[4818]: I1203 06:50:18.961749 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-config\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:18 crc kubenswrapper[4818]: I1203 06:50:18.961856 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:18 crc kubenswrapper[4818]: I1203 06:50:18.961920 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:18 crc kubenswrapper[4818]: I1203 06:50:18.961965 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:19 crc kubenswrapper[4818]: I1203 06:50:19.063770 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:19 crc kubenswrapper[4818]: I1203 06:50:19.064317 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk54h\" (UniqueName: \"kubernetes.io/projected/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-kube-api-access-lk54h\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:19 crc kubenswrapper[4818]: I1203 06:50:19.064352 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:19 crc kubenswrapper[4818]: I1203 06:50:19.064401 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-config\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:19 crc kubenswrapper[4818]: I1203 06:50:19.064476 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:19 crc kubenswrapper[4818]: I1203 06:50:19.064524 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:19 crc kubenswrapper[4818]: I1203 06:50:19.064556 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:19 crc kubenswrapper[4818]: I1203 06:50:19.065404 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:19 crc kubenswrapper[4818]: I1203 06:50:19.065478 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:19 crc kubenswrapper[4818]: I1203 06:50:19.065629 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:19 crc kubenswrapper[4818]: I1203 06:50:19.065742 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:19 crc kubenswrapper[4818]: I1203 06:50:19.066087 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:19 crc kubenswrapper[4818]: I1203 06:50:19.066223 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-config\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:19 crc kubenswrapper[4818]: I1203 06:50:19.081991 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk54h\" (UniqueName: \"kubernetes.io/projected/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-kube-api-access-lk54h\") pod \"dnsmasq-dns-79bd4cc8c9-lgdlb\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:19 crc kubenswrapper[4818]: I1203 06:50:19.177033 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:19 crc kubenswrapper[4818]: I1203 06:50:19.684047 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-lgdlb"] Dec 03 06:50:19 crc kubenswrapper[4818]: I1203 06:50:19.902797 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" event={"ID":"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6","Type":"ContainerStarted","Data":"84122ffb5cd9912332280af638d663ab362acf444622c8912a59f340e2b43ac4"} Dec 03 06:50:20 crc kubenswrapper[4818]: I1203 06:50:20.916362 4818 generic.go:334] "Generic (PLEG): container finished" podID="d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6" containerID="dc96f106a6aeafa9e9c38794ff5fd649a2d5931403a0c14c13aeec09a9547bb4" exitCode=0 Dec 03 06:50:20 crc kubenswrapper[4818]: I1203 06:50:20.916453 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" event={"ID":"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6","Type":"ContainerDied","Data":"dc96f106a6aeafa9e9c38794ff5fd649a2d5931403a0c14c13aeec09a9547bb4"} Dec 03 06:50:20 crc kubenswrapper[4818]: I1203 06:50:20.923664 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8078039a-8674-40c1-ba1a-ff8244d17e16","Type":"ContainerStarted","Data":"b85c6681db090ec115d2794f26dcb1afc5ecf1d1656298aa368002a2433b1277"} Dec 03 06:50:21 crc kubenswrapper[4818]: I1203 06:50:21.942805 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" event={"ID":"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6","Type":"ContainerStarted","Data":"bf8b9ede7f26ca65c2b87e98e00605fac03b493e1c98acedd7e01662c2b0e7b7"} Dec 03 06:50:21 crc kubenswrapper[4818]: I1203 06:50:21.980807 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" podStartSLOduration=3.9807794469999997 podStartE2EDuration="3.980779447s" podCreationTimestamp="2025-12-03 06:50:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:50:21.966797012 +0000 UTC m=+1379.658405804" watchObservedRunningTime="2025-12-03 06:50:21.980779447 +0000 UTC m=+1379.672388239" Dec 03 06:50:22 crc kubenswrapper[4818]: I1203 06:50:22.953866 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.179051 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.263705 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bsnnk"] Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.264101 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" podUID="0b06c002-0804-463d-84ca-0cc7fa8453b8" containerName="dnsmasq-dns" containerID="cri-o://3077c286134324a6da59bb654d265558445cd43dc0afd601b184056ae709752b" gracePeriod=10 Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.464006 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55478c4467-gbkbl"] Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.465898 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.487370 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-gbkbl"] Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.487852 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.487896 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.487933 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.488057 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpwzf\" (UniqueName: \"kubernetes.io/projected/19b91d65-7f62-47e7-b12c-33135811b76b-kube-api-access-kpwzf\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.488354 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-dns-svc\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.488455 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-config\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.488497 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.589383 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.589456 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.589478 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.589509 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.589537 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpwzf\" (UniqueName: \"kubernetes.io/projected/19b91d65-7f62-47e7-b12c-33135811b76b-kube-api-access-kpwzf\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.589603 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-dns-svc\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.589642 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-config\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.590550 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.590597 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.590803 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-config\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.590805 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.591055 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-dns-svc\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.591332 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/19b91d65-7f62-47e7-b12c-33135811b76b-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.609050 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpwzf\" (UniqueName: \"kubernetes.io/projected/19b91d65-7f62-47e7-b12c-33135811b76b-kube-api-access-kpwzf\") pod \"dnsmasq-dns-55478c4467-gbkbl\" (UID: \"19b91d65-7f62-47e7-b12c-33135811b76b\") " pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.803737 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.924065 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.997389 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g2mw\" (UniqueName: \"kubernetes.io/projected/0b06c002-0804-463d-84ca-0cc7fa8453b8-kube-api-access-6g2mw\") pod \"0b06c002-0804-463d-84ca-0cc7fa8453b8\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.997438 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-ovsdbserver-sb\") pod \"0b06c002-0804-463d-84ca-0cc7fa8453b8\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.997585 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-dns-svc\") pod \"0b06c002-0804-463d-84ca-0cc7fa8453b8\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.997773 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-dns-swift-storage-0\") pod \"0b06c002-0804-463d-84ca-0cc7fa8453b8\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.997852 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-ovsdbserver-nb\") pod \"0b06c002-0804-463d-84ca-0cc7fa8453b8\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " Dec 03 06:50:29 crc kubenswrapper[4818]: I1203 06:50:29.997902 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-config\") pod \"0b06c002-0804-463d-84ca-0cc7fa8453b8\" (UID: \"0b06c002-0804-463d-84ca-0cc7fa8453b8\") " Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.004874 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b06c002-0804-463d-84ca-0cc7fa8453b8-kube-api-access-6g2mw" (OuterVolumeSpecName: "kube-api-access-6g2mw") pod "0b06c002-0804-463d-84ca-0cc7fa8453b8" (UID: "0b06c002-0804-463d-84ca-0cc7fa8453b8"). InnerVolumeSpecName "kube-api-access-6g2mw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.039547 4818 generic.go:334] "Generic (PLEG): container finished" podID="0b06c002-0804-463d-84ca-0cc7fa8453b8" containerID="3077c286134324a6da59bb654d265558445cd43dc0afd601b184056ae709752b" exitCode=0 Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.039594 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" event={"ID":"0b06c002-0804-463d-84ca-0cc7fa8453b8","Type":"ContainerDied","Data":"3077c286134324a6da59bb654d265558445cd43dc0afd601b184056ae709752b"} Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.039621 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" event={"ID":"0b06c002-0804-463d-84ca-0cc7fa8453b8","Type":"ContainerDied","Data":"a447fd0071848427583a78d064adcc0467026738f3279f409b2378b320625636"} Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.039637 4818 scope.go:117] "RemoveContainer" containerID="3077c286134324a6da59bb654d265558445cd43dc0afd601b184056ae709752b" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.039763 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-bsnnk" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.053559 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0b06c002-0804-463d-84ca-0cc7fa8453b8" (UID: "0b06c002-0804-463d-84ca-0cc7fa8453b8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.056418 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0b06c002-0804-463d-84ca-0cc7fa8453b8" (UID: "0b06c002-0804-463d-84ca-0cc7fa8453b8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.068257 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0b06c002-0804-463d-84ca-0cc7fa8453b8" (UID: "0b06c002-0804-463d-84ca-0cc7fa8453b8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.070843 4818 scope.go:117] "RemoveContainer" containerID="e484eb2f82618daab92823e31d09951aa822a64bcf65939bc3dfb37d798dc5a6" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.073763 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0b06c002-0804-463d-84ca-0cc7fa8453b8" (UID: "0b06c002-0804-463d-84ca-0cc7fa8453b8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.075265 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-config" (OuterVolumeSpecName: "config") pod "0b06c002-0804-463d-84ca-0cc7fa8453b8" (UID: "0b06c002-0804-463d-84ca-0cc7fa8453b8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.094788 4818 scope.go:117] "RemoveContainer" containerID="3077c286134324a6da59bb654d265558445cd43dc0afd601b184056ae709752b" Dec 03 06:50:30 crc kubenswrapper[4818]: E1203 06:50:30.095212 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3077c286134324a6da59bb654d265558445cd43dc0afd601b184056ae709752b\": container with ID starting with 3077c286134324a6da59bb654d265558445cd43dc0afd601b184056ae709752b not found: ID does not exist" containerID="3077c286134324a6da59bb654d265558445cd43dc0afd601b184056ae709752b" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.095263 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3077c286134324a6da59bb654d265558445cd43dc0afd601b184056ae709752b"} err="failed to get container status \"3077c286134324a6da59bb654d265558445cd43dc0afd601b184056ae709752b\": rpc error: code = NotFound desc = could not find container \"3077c286134324a6da59bb654d265558445cd43dc0afd601b184056ae709752b\": container with ID starting with 3077c286134324a6da59bb654d265558445cd43dc0afd601b184056ae709752b not found: ID does not exist" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.095281 4818 scope.go:117] "RemoveContainer" containerID="e484eb2f82618daab92823e31d09951aa822a64bcf65939bc3dfb37d798dc5a6" Dec 03 06:50:30 crc kubenswrapper[4818]: E1203 06:50:30.095678 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e484eb2f82618daab92823e31d09951aa822a64bcf65939bc3dfb37d798dc5a6\": container with ID starting with e484eb2f82618daab92823e31d09951aa822a64bcf65939bc3dfb37d798dc5a6 not found: ID does not exist" containerID="e484eb2f82618daab92823e31d09951aa822a64bcf65939bc3dfb37d798dc5a6" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.095721 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e484eb2f82618daab92823e31d09951aa822a64bcf65939bc3dfb37d798dc5a6"} err="failed to get container status \"e484eb2f82618daab92823e31d09951aa822a64bcf65939bc3dfb37d798dc5a6\": rpc error: code = NotFound desc = could not find container \"e484eb2f82618daab92823e31d09951aa822a64bcf65939bc3dfb37d798dc5a6\": container with ID starting with e484eb2f82618daab92823e31d09951aa822a64bcf65939bc3dfb37d798dc5a6 not found: ID does not exist" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.100385 4818 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.100415 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.100425 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.100435 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g2mw\" (UniqueName: \"kubernetes.io/projected/0b06c002-0804-463d-84ca-0cc7fa8453b8-kube-api-access-6g2mw\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.100445 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.100453 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b06c002-0804-463d-84ca-0cc7fa8453b8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.273925 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-gbkbl"] Dec 03 06:50:30 crc kubenswrapper[4818]: W1203 06:50:30.277132 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19b91d65_7f62_47e7_b12c_33135811b76b.slice/crio-83f5db361662baeb3c886c656a7d6a0c6ee8855124bdcb8cfdc25c2defd83aa8 WatchSource:0}: Error finding container 83f5db361662baeb3c886c656a7d6a0c6ee8855124bdcb8cfdc25c2defd83aa8: Status 404 returned error can't find the container with id 83f5db361662baeb3c886c656a7d6a0c6ee8855124bdcb8cfdc25c2defd83aa8 Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.376518 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bsnnk"] Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.384523 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bsnnk"] Dec 03 06:50:30 crc kubenswrapper[4818]: I1203 06:50:30.748930 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b06c002-0804-463d-84ca-0cc7fa8453b8" path="/var/lib/kubelet/pods/0b06c002-0804-463d-84ca-0cc7fa8453b8/volumes" Dec 03 06:50:31 crc kubenswrapper[4818]: I1203 06:50:31.049956 4818 generic.go:334] "Generic (PLEG): container finished" podID="19b91d65-7f62-47e7-b12c-33135811b76b" containerID="64564d2bfd4a945359393f8798b2ba7487eedd4b735d3cd23e91561e00a24228" exitCode=0 Dec 03 06:50:31 crc kubenswrapper[4818]: I1203 06:50:31.050069 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-gbkbl" event={"ID":"19b91d65-7f62-47e7-b12c-33135811b76b","Type":"ContainerDied","Data":"64564d2bfd4a945359393f8798b2ba7487eedd4b735d3cd23e91561e00a24228"} Dec 03 06:50:31 crc kubenswrapper[4818]: I1203 06:50:31.050274 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-gbkbl" event={"ID":"19b91d65-7f62-47e7-b12c-33135811b76b","Type":"ContainerStarted","Data":"83f5db361662baeb3c886c656a7d6a0c6ee8855124bdcb8cfdc25c2defd83aa8"} Dec 03 06:50:32 crc kubenswrapper[4818]: I1203 06:50:32.063893 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-gbkbl" event={"ID":"19b91d65-7f62-47e7-b12c-33135811b76b","Type":"ContainerStarted","Data":"5a2e4719e4e0ff8df794fa12ff231c354fe4cc2275f13c2fe6b9d2734fedc998"} Dec 03 06:50:32 crc kubenswrapper[4818]: I1203 06:50:32.065190 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:32 crc kubenswrapper[4818]: I1203 06:50:32.091157 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55478c4467-gbkbl" podStartSLOduration=3.09113879 podStartE2EDuration="3.09113879s" podCreationTimestamp="2025-12-03 06:50:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:50:32.083516552 +0000 UTC m=+1389.775125314" watchObservedRunningTime="2025-12-03 06:50:32.09113879 +0000 UTC m=+1389.782747532" Dec 03 06:50:39 crc kubenswrapper[4818]: I1203 06:50:39.805016 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55478c4467-gbkbl" Dec 03 06:50:39 crc kubenswrapper[4818]: I1203 06:50:39.895348 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-lgdlb"] Dec 03 06:50:39 crc kubenswrapper[4818]: I1203 06:50:39.895646 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" podUID="d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6" containerName="dnsmasq-dns" containerID="cri-o://bf8b9ede7f26ca65c2b87e98e00605fac03b493e1c98acedd7e01662c2b0e7b7" gracePeriod=10 Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.170977 4818 generic.go:334] "Generic (PLEG): container finished" podID="d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6" containerID="bf8b9ede7f26ca65c2b87e98e00605fac03b493e1c98acedd7e01662c2b0e7b7" exitCode=0 Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.171003 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" event={"ID":"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6","Type":"ContainerDied","Data":"bf8b9ede7f26ca65c2b87e98e00605fac03b493e1c98acedd7e01662c2b0e7b7"} Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.376380 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.500949 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-openstack-edpm-ipam\") pod \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.501003 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-config\") pod \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.501040 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk54h\" (UniqueName: \"kubernetes.io/projected/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-kube-api-access-lk54h\") pod \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.501200 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-dns-swift-storage-0\") pod \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.501221 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-dns-svc\") pod \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.501246 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-ovsdbserver-sb\") pod \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.501271 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-ovsdbserver-nb\") pod \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\" (UID: \"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6\") " Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.507357 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-kube-api-access-lk54h" (OuterVolumeSpecName: "kube-api-access-lk54h") pod "d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6" (UID: "d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6"). InnerVolumeSpecName "kube-api-access-lk54h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.554186 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6" (UID: "d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.558419 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6" (UID: "d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.558862 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6" (UID: "d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.568039 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6" (UID: "d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.580187 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-config" (OuterVolumeSpecName: "config") pod "d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6" (UID: "d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.590166 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6" (UID: "d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.603355 4818 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-config\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.603391 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk54h\" (UniqueName: \"kubernetes.io/projected/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-kube-api-access-lk54h\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.603402 4818 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.603412 4818 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.603421 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.603429 4818 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:40 crc kubenswrapper[4818]: I1203 06:50:40.603437 4818 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 06:50:41 crc kubenswrapper[4818]: I1203 06:50:41.182768 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" event={"ID":"d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6","Type":"ContainerDied","Data":"84122ffb5cd9912332280af638d663ab362acf444622c8912a59f340e2b43ac4"} Dec 03 06:50:41 crc kubenswrapper[4818]: I1203 06:50:41.182855 4818 scope.go:117] "RemoveContainer" containerID="bf8b9ede7f26ca65c2b87e98e00605fac03b493e1c98acedd7e01662c2b0e7b7" Dec 03 06:50:41 crc kubenswrapper[4818]: I1203 06:50:41.182890 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-lgdlb" Dec 03 06:50:41 crc kubenswrapper[4818]: I1203 06:50:41.210428 4818 scope.go:117] "RemoveContainer" containerID="dc96f106a6aeafa9e9c38794ff5fd649a2d5931403a0c14c13aeec09a9547bb4" Dec 03 06:50:41 crc kubenswrapper[4818]: I1203 06:50:41.215585 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-lgdlb"] Dec 03 06:50:41 crc kubenswrapper[4818]: I1203 06:50:41.229292 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-lgdlb"] Dec 03 06:50:42 crc kubenswrapper[4818]: I1203 06:50:42.756608 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6" path="/var/lib/kubelet/pods/d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6/volumes" Dec 03 06:50:50 crc kubenswrapper[4818]: I1203 06:50:50.936467 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x5c96"] Dec 03 06:50:50 crc kubenswrapper[4818]: E1203 06:50:50.937485 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6" containerName="init" Dec 03 06:50:50 crc kubenswrapper[4818]: I1203 06:50:50.937501 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6" containerName="init" Dec 03 06:50:50 crc kubenswrapper[4818]: E1203 06:50:50.937520 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b06c002-0804-463d-84ca-0cc7fa8453b8" containerName="dnsmasq-dns" Dec 03 06:50:50 crc kubenswrapper[4818]: I1203 06:50:50.937527 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b06c002-0804-463d-84ca-0cc7fa8453b8" containerName="dnsmasq-dns" Dec 03 06:50:50 crc kubenswrapper[4818]: E1203 06:50:50.937548 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b06c002-0804-463d-84ca-0cc7fa8453b8" containerName="init" Dec 03 06:50:50 crc kubenswrapper[4818]: I1203 06:50:50.937555 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b06c002-0804-463d-84ca-0cc7fa8453b8" containerName="init" Dec 03 06:50:50 crc kubenswrapper[4818]: E1203 06:50:50.937581 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6" containerName="dnsmasq-dns" Dec 03 06:50:50 crc kubenswrapper[4818]: I1203 06:50:50.937588 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6" containerName="dnsmasq-dns" Dec 03 06:50:50 crc kubenswrapper[4818]: I1203 06:50:50.937801 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4e71549-2ae5-4f9d-8ffe-b8a74651f8a6" containerName="dnsmasq-dns" Dec 03 06:50:50 crc kubenswrapper[4818]: I1203 06:50:50.937842 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b06c002-0804-463d-84ca-0cc7fa8453b8" containerName="dnsmasq-dns" Dec 03 06:50:50 crc kubenswrapper[4818]: I1203 06:50:50.939242 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5c96" Dec 03 06:50:50 crc kubenswrapper[4818]: I1203 06:50:50.957405 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x5c96"] Dec 03 06:50:51 crc kubenswrapper[4818]: I1203 06:50:51.111165 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7glsm\" (UniqueName: \"kubernetes.io/projected/4002e49a-cb3f-477f-bb28-db0ddbc15230-kube-api-access-7glsm\") pod \"redhat-operators-x5c96\" (UID: \"4002e49a-cb3f-477f-bb28-db0ddbc15230\") " pod="openshift-marketplace/redhat-operators-x5c96" Dec 03 06:50:51 crc kubenswrapper[4818]: I1203 06:50:51.111617 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4002e49a-cb3f-477f-bb28-db0ddbc15230-utilities\") pod \"redhat-operators-x5c96\" (UID: \"4002e49a-cb3f-477f-bb28-db0ddbc15230\") " pod="openshift-marketplace/redhat-operators-x5c96" Dec 03 06:50:51 crc kubenswrapper[4818]: I1203 06:50:51.111782 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4002e49a-cb3f-477f-bb28-db0ddbc15230-catalog-content\") pod \"redhat-operators-x5c96\" (UID: \"4002e49a-cb3f-477f-bb28-db0ddbc15230\") " pod="openshift-marketplace/redhat-operators-x5c96" Dec 03 06:50:51 crc kubenswrapper[4818]: I1203 06:50:51.212928 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7glsm\" (UniqueName: \"kubernetes.io/projected/4002e49a-cb3f-477f-bb28-db0ddbc15230-kube-api-access-7glsm\") pod \"redhat-operators-x5c96\" (UID: \"4002e49a-cb3f-477f-bb28-db0ddbc15230\") " pod="openshift-marketplace/redhat-operators-x5c96" Dec 03 06:50:51 crc kubenswrapper[4818]: I1203 06:50:51.213059 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4002e49a-cb3f-477f-bb28-db0ddbc15230-utilities\") pod \"redhat-operators-x5c96\" (UID: \"4002e49a-cb3f-477f-bb28-db0ddbc15230\") " pod="openshift-marketplace/redhat-operators-x5c96" Dec 03 06:50:51 crc kubenswrapper[4818]: I1203 06:50:51.213106 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4002e49a-cb3f-477f-bb28-db0ddbc15230-catalog-content\") pod \"redhat-operators-x5c96\" (UID: \"4002e49a-cb3f-477f-bb28-db0ddbc15230\") " pod="openshift-marketplace/redhat-operators-x5c96" Dec 03 06:50:51 crc kubenswrapper[4818]: I1203 06:50:51.213638 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4002e49a-cb3f-477f-bb28-db0ddbc15230-utilities\") pod \"redhat-operators-x5c96\" (UID: \"4002e49a-cb3f-477f-bb28-db0ddbc15230\") " pod="openshift-marketplace/redhat-operators-x5c96" Dec 03 06:50:51 crc kubenswrapper[4818]: I1203 06:50:51.213665 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4002e49a-cb3f-477f-bb28-db0ddbc15230-catalog-content\") pod \"redhat-operators-x5c96\" (UID: \"4002e49a-cb3f-477f-bb28-db0ddbc15230\") " pod="openshift-marketplace/redhat-operators-x5c96" Dec 03 06:50:51 crc kubenswrapper[4818]: I1203 06:50:51.234415 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7glsm\" (UniqueName: \"kubernetes.io/projected/4002e49a-cb3f-477f-bb28-db0ddbc15230-kube-api-access-7glsm\") pod \"redhat-operators-x5c96\" (UID: \"4002e49a-cb3f-477f-bb28-db0ddbc15230\") " pod="openshift-marketplace/redhat-operators-x5c96" Dec 03 06:50:51 crc kubenswrapper[4818]: I1203 06:50:51.260072 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5c96" Dec 03 06:50:51 crc kubenswrapper[4818]: I1203 06:50:51.306591 4818 generic.go:334] "Generic (PLEG): container finished" podID="e6fa481e-32ce-4840-b457-0a692558cdb5" containerID="1bed1e3c0e6ad25d79ffed0a1e1704cf610cbc76ef15c729896c79167c93693f" exitCode=0 Dec 03 06:50:51 crc kubenswrapper[4818]: I1203 06:50:51.306664 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e6fa481e-32ce-4840-b457-0a692558cdb5","Type":"ContainerDied","Data":"1bed1e3c0e6ad25d79ffed0a1e1704cf610cbc76ef15c729896c79167c93693f"} Dec 03 06:50:51 crc kubenswrapper[4818]: I1203 06:50:51.785745 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x5c96"] Dec 03 06:50:52 crc kubenswrapper[4818]: I1203 06:50:52.315856 4818 generic.go:334] "Generic (PLEG): container finished" podID="4002e49a-cb3f-477f-bb28-db0ddbc15230" containerID="5615f5b200ddafc1778b890da77f0f90a388c72b7f1d22babcafb5ba6d5a530b" exitCode=0 Dec 03 06:50:52 crc kubenswrapper[4818]: I1203 06:50:52.315946 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5c96" event={"ID":"4002e49a-cb3f-477f-bb28-db0ddbc15230","Type":"ContainerDied","Data":"5615f5b200ddafc1778b890da77f0f90a388c72b7f1d22babcafb5ba6d5a530b"} Dec 03 06:50:52 crc kubenswrapper[4818]: I1203 06:50:52.316229 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5c96" event={"ID":"4002e49a-cb3f-477f-bb28-db0ddbc15230","Type":"ContainerStarted","Data":"5e626b69cfff97add226a65f3d2bfda160c575ecb03e08d7bf33016ef9e2ea1c"} Dec 03 06:50:52 crc kubenswrapper[4818]: I1203 06:50:52.325875 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e6fa481e-32ce-4840-b457-0a692558cdb5","Type":"ContainerStarted","Data":"1f97ba3f4165d5f386fe1af0d3ae2a4c326bd9b3fc80bbbcbb7b57ff92399ad2"} Dec 03 06:50:52 crc kubenswrapper[4818]: I1203 06:50:52.326105 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 06:50:52 crc kubenswrapper[4818]: I1203 06:50:52.904720 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.904701708 podStartE2EDuration="37.904701708s" podCreationTimestamp="2025-12-03 06:50:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:50:52.430890711 +0000 UTC m=+1410.122499473" watchObservedRunningTime="2025-12-03 06:50:52.904701708 +0000 UTC m=+1410.596310460" Dec 03 06:50:52 crc kubenswrapper[4818]: I1203 06:50:52.905221 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh"] Dec 03 06:50:52 crc kubenswrapper[4818]: I1203 06:50:52.906429 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" Dec 03 06:50:52 crc kubenswrapper[4818]: I1203 06:50:52.909329 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2bpsv" Dec 03 06:50:52 crc kubenswrapper[4818]: I1203 06:50:52.909520 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 06:50:52 crc kubenswrapper[4818]: I1203 06:50:52.909557 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 06:50:52 crc kubenswrapper[4818]: I1203 06:50:52.909655 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 06:50:52 crc kubenswrapper[4818]: I1203 06:50:52.921807 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh"] Dec 03 06:50:53 crc kubenswrapper[4818]: I1203 06:50:53.067058 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh\" (UID: \"37db016a-a063-4751-bf39-aadd6746fa13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" Dec 03 06:50:53 crc kubenswrapper[4818]: I1203 06:50:53.067127 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh\" (UID: \"37db016a-a063-4751-bf39-aadd6746fa13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" Dec 03 06:50:53 crc kubenswrapper[4818]: I1203 06:50:53.067441 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh\" (UID: \"37db016a-a063-4751-bf39-aadd6746fa13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" Dec 03 06:50:53 crc kubenswrapper[4818]: I1203 06:50:53.067639 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjgqd\" (UniqueName: \"kubernetes.io/projected/37db016a-a063-4751-bf39-aadd6746fa13-kube-api-access-mjgqd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh\" (UID: \"37db016a-a063-4751-bf39-aadd6746fa13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" Dec 03 06:50:53 crc kubenswrapper[4818]: I1203 06:50:53.169950 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh\" (UID: \"37db016a-a063-4751-bf39-aadd6746fa13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" Dec 03 06:50:53 crc kubenswrapper[4818]: I1203 06:50:53.170120 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjgqd\" (UniqueName: \"kubernetes.io/projected/37db016a-a063-4751-bf39-aadd6746fa13-kube-api-access-mjgqd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh\" (UID: \"37db016a-a063-4751-bf39-aadd6746fa13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" Dec 03 06:50:53 crc kubenswrapper[4818]: I1203 06:50:53.170280 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh\" (UID: \"37db016a-a063-4751-bf39-aadd6746fa13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" Dec 03 06:50:53 crc kubenswrapper[4818]: I1203 06:50:53.170334 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh\" (UID: \"37db016a-a063-4751-bf39-aadd6746fa13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" Dec 03 06:50:53 crc kubenswrapper[4818]: I1203 06:50:53.177410 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh\" (UID: \"37db016a-a063-4751-bf39-aadd6746fa13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" Dec 03 06:50:53 crc kubenswrapper[4818]: I1203 06:50:53.177473 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh\" (UID: \"37db016a-a063-4751-bf39-aadd6746fa13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" Dec 03 06:50:53 crc kubenswrapper[4818]: I1203 06:50:53.181979 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh\" (UID: \"37db016a-a063-4751-bf39-aadd6746fa13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" Dec 03 06:50:53 crc kubenswrapper[4818]: I1203 06:50:53.202175 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjgqd\" (UniqueName: \"kubernetes.io/projected/37db016a-a063-4751-bf39-aadd6746fa13-kube-api-access-mjgqd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh\" (UID: \"37db016a-a063-4751-bf39-aadd6746fa13\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" Dec 03 06:50:53 crc kubenswrapper[4818]: I1203 06:50:53.240575 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" Dec 03 06:50:53 crc kubenswrapper[4818]: I1203 06:50:53.354967 4818 generic.go:334] "Generic (PLEG): container finished" podID="8078039a-8674-40c1-ba1a-ff8244d17e16" containerID="b85c6681db090ec115d2794f26dcb1afc5ecf1d1656298aa368002a2433b1277" exitCode=0 Dec 03 06:50:53 crc kubenswrapper[4818]: I1203 06:50:53.355475 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8078039a-8674-40c1-ba1a-ff8244d17e16","Type":"ContainerDied","Data":"b85c6681db090ec115d2794f26dcb1afc5ecf1d1656298aa368002a2433b1277"} Dec 03 06:50:53 crc kubenswrapper[4818]: I1203 06:50:53.381211 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5c96" event={"ID":"4002e49a-cb3f-477f-bb28-db0ddbc15230","Type":"ContainerStarted","Data":"c2fe0f3717047ec1cc4f7793d8718ac0d1977d617ab24837256f54cf7be299a4"} Dec 03 06:50:54 crc kubenswrapper[4818]: I1203 06:50:54.529507 4818 generic.go:334] "Generic (PLEG): container finished" podID="4002e49a-cb3f-477f-bb28-db0ddbc15230" containerID="c2fe0f3717047ec1cc4f7793d8718ac0d1977d617ab24837256f54cf7be299a4" exitCode=0 Dec 03 06:50:54 crc kubenswrapper[4818]: I1203 06:50:54.530853 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5c96" event={"ID":"4002e49a-cb3f-477f-bb28-db0ddbc15230","Type":"ContainerDied","Data":"c2fe0f3717047ec1cc4f7793d8718ac0d1977d617ab24837256f54cf7be299a4"} Dec 03 06:50:54 crc kubenswrapper[4818]: I1203 06:50:54.618692 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh"] Dec 03 06:50:55 crc kubenswrapper[4818]: I1203 06:50:55.545711 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8078039a-8674-40c1-ba1a-ff8244d17e16","Type":"ContainerStarted","Data":"348c40eecb4fc1884927a74990a99f2ad5611871abd848dfa1b85f193fa40d7d"} Dec 03 06:50:55 crc kubenswrapper[4818]: I1203 06:50:55.547066 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:50:55 crc kubenswrapper[4818]: I1203 06:50:55.549452 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5c96" event={"ID":"4002e49a-cb3f-477f-bb28-db0ddbc15230","Type":"ContainerStarted","Data":"8825dd7dfdbc375d100989a49cccf84cb650673fd46e63d97766601ae8f5a413"} Dec 03 06:50:55 crc kubenswrapper[4818]: I1203 06:50:55.551098 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" event={"ID":"37db016a-a063-4751-bf39-aadd6746fa13","Type":"ContainerStarted","Data":"e6c3fc1fe24a9bd2ee59b13b33a6a287394b559462143e43c35c7d5a6978b215"} Dec 03 06:50:55 crc kubenswrapper[4818]: I1203 06:50:55.589584 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.589559877 podStartE2EDuration="38.589559877s" podCreationTimestamp="2025-12-03 06:50:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 06:50:55.572316071 +0000 UTC m=+1413.263924823" watchObservedRunningTime="2025-12-03 06:50:55.589559877 +0000 UTC m=+1413.281168629" Dec 03 06:50:55 crc kubenswrapper[4818]: I1203 06:50:55.598635 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x5c96" podStartSLOduration=2.73681016 podStartE2EDuration="5.598613411s" podCreationTimestamp="2025-12-03 06:50:50 +0000 UTC" firstStartedPulling="2025-12-03 06:50:52.317556811 +0000 UTC m=+1410.009165563" lastFinishedPulling="2025-12-03 06:50:55.179360062 +0000 UTC m=+1412.870968814" observedRunningTime="2025-12-03 06:50:55.590432509 +0000 UTC m=+1413.282041261" watchObservedRunningTime="2025-12-03 06:50:55.598613411 +0000 UTC m=+1413.290222163" Dec 03 06:51:01 crc kubenswrapper[4818]: I1203 06:51:01.271996 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x5c96" Dec 03 06:51:01 crc kubenswrapper[4818]: I1203 06:51:01.272575 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x5c96" Dec 03 06:51:02 crc kubenswrapper[4818]: I1203 06:51:02.341607 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x5c96" podUID="4002e49a-cb3f-477f-bb28-db0ddbc15230" containerName="registry-server" probeResult="failure" output=< Dec 03 06:51:02 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Dec 03 06:51:02 crc kubenswrapper[4818]: > Dec 03 06:51:06 crc kubenswrapper[4818]: I1203 06:51:06.183016 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 06:51:06 crc kubenswrapper[4818]: I1203 06:51:06.665171 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" event={"ID":"37db016a-a063-4751-bf39-aadd6746fa13","Type":"ContainerStarted","Data":"3e15df8d2799fe0c8cf5a796bb8ed6f9f3f63e877873edc819f6a758af02a7a3"} Dec 03 06:51:06 crc kubenswrapper[4818]: I1203 06:51:06.689069 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" podStartSLOduration=3.558371717 podStartE2EDuration="14.68904633s" podCreationTimestamp="2025-12-03 06:50:52 +0000 UTC" firstStartedPulling="2025-12-03 06:50:54.619074108 +0000 UTC m=+1412.310682860" lastFinishedPulling="2025-12-03 06:51:05.749748721 +0000 UTC m=+1423.441357473" observedRunningTime="2025-12-03 06:51:06.686858126 +0000 UTC m=+1424.378466888" watchObservedRunningTime="2025-12-03 06:51:06.68904633 +0000 UTC m=+1424.380655082" Dec 03 06:51:07 crc kubenswrapper[4818]: I1203 06:51:07.825024 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 06:51:12 crc kubenswrapper[4818]: I1203 06:51:12.307837 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x5c96" podUID="4002e49a-cb3f-477f-bb28-db0ddbc15230" containerName="registry-server" probeResult="failure" output=< Dec 03 06:51:12 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Dec 03 06:51:12 crc kubenswrapper[4818]: > Dec 03 06:51:17 crc kubenswrapper[4818]: I1203 06:51:17.758416 4818 generic.go:334] "Generic (PLEG): container finished" podID="37db016a-a063-4751-bf39-aadd6746fa13" containerID="3e15df8d2799fe0c8cf5a796bb8ed6f9f3f63e877873edc819f6a758af02a7a3" exitCode=0 Dec 03 06:51:17 crc kubenswrapper[4818]: I1203 06:51:17.758497 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" event={"ID":"37db016a-a063-4751-bf39-aadd6746fa13","Type":"ContainerDied","Data":"3e15df8d2799fe0c8cf5a796bb8ed6f9f3f63e877873edc819f6a758af02a7a3"} Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.176223 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.266384 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjgqd\" (UniqueName: \"kubernetes.io/projected/37db016a-a063-4751-bf39-aadd6746fa13-kube-api-access-mjgqd\") pod \"37db016a-a063-4751-bf39-aadd6746fa13\" (UID: \"37db016a-a063-4751-bf39-aadd6746fa13\") " Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.266523 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-inventory\") pod \"37db016a-a063-4751-bf39-aadd6746fa13\" (UID: \"37db016a-a063-4751-bf39-aadd6746fa13\") " Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.266583 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-ssh-key\") pod \"37db016a-a063-4751-bf39-aadd6746fa13\" (UID: \"37db016a-a063-4751-bf39-aadd6746fa13\") " Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.266636 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-repo-setup-combined-ca-bundle\") pod \"37db016a-a063-4751-bf39-aadd6746fa13\" (UID: \"37db016a-a063-4751-bf39-aadd6746fa13\") " Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.272993 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37db016a-a063-4751-bf39-aadd6746fa13-kube-api-access-mjgqd" (OuterVolumeSpecName: "kube-api-access-mjgqd") pod "37db016a-a063-4751-bf39-aadd6746fa13" (UID: "37db016a-a063-4751-bf39-aadd6746fa13"). InnerVolumeSpecName "kube-api-access-mjgqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.282082 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "37db016a-a063-4751-bf39-aadd6746fa13" (UID: "37db016a-a063-4751-bf39-aadd6746fa13"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.295440 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-inventory" (OuterVolumeSpecName: "inventory") pod "37db016a-a063-4751-bf39-aadd6746fa13" (UID: "37db016a-a063-4751-bf39-aadd6746fa13"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.304075 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "37db016a-a063-4751-bf39-aadd6746fa13" (UID: "37db016a-a063-4751-bf39-aadd6746fa13"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.369233 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjgqd\" (UniqueName: \"kubernetes.io/projected/37db016a-a063-4751-bf39-aadd6746fa13-kube-api-access-mjgqd\") on node \"crc\" DevicePath \"\"" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.369265 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.369278 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.369290 4818 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37db016a-a063-4751-bf39-aadd6746fa13-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.781115 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" event={"ID":"37db016a-a063-4751-bf39-aadd6746fa13","Type":"ContainerDied","Data":"e6c3fc1fe24a9bd2ee59b13b33a6a287394b559462143e43c35c7d5a6978b215"} Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.781152 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6c3fc1fe24a9bd2ee59b13b33a6a287394b559462143e43c35c7d5a6978b215" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.781258 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.865601 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q"] Dec 03 06:51:19 crc kubenswrapper[4818]: E1203 06:51:19.865996 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37db016a-a063-4751-bf39-aadd6746fa13" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.866010 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="37db016a-a063-4751-bf39-aadd6746fa13" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.866214 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="37db016a-a063-4751-bf39-aadd6746fa13" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.866904 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.869645 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2bpsv" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.869776 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.870414 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.871280 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 06:51:19 crc kubenswrapper[4818]: I1203 06:51:19.877467 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q"] Dec 03 06:51:20 crc kubenswrapper[4818]: I1203 06:51:20.013153 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nsd6q\" (UID: \"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" Dec 03 06:51:20 crc kubenswrapper[4818]: I1203 06:51:20.013542 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nsd6q\" (UID: \"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" Dec 03 06:51:20 crc kubenswrapper[4818]: I1203 06:51:20.013665 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnj96\" (UniqueName: \"kubernetes.io/projected/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-kube-api-access-dnj96\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nsd6q\" (UID: \"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" Dec 03 06:51:20 crc kubenswrapper[4818]: I1203 06:51:20.115334 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nsd6q\" (UID: \"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" Dec 03 06:51:20 crc kubenswrapper[4818]: I1203 06:51:20.115957 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnj96\" (UniqueName: \"kubernetes.io/projected/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-kube-api-access-dnj96\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nsd6q\" (UID: \"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" Dec 03 06:51:20 crc kubenswrapper[4818]: I1203 06:51:20.116088 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nsd6q\" (UID: \"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" Dec 03 06:51:20 crc kubenswrapper[4818]: I1203 06:51:20.121293 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nsd6q\" (UID: \"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" Dec 03 06:51:20 crc kubenswrapper[4818]: I1203 06:51:20.121955 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nsd6q\" (UID: \"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" Dec 03 06:51:20 crc kubenswrapper[4818]: I1203 06:51:20.133981 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnj96\" (UniqueName: \"kubernetes.io/projected/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-kube-api-access-dnj96\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-nsd6q\" (UID: \"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" Dec 03 06:51:20 crc kubenswrapper[4818]: I1203 06:51:20.236214 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" Dec 03 06:51:20 crc kubenswrapper[4818]: I1203 06:51:20.735982 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q"] Dec 03 06:51:20 crc kubenswrapper[4818]: I1203 06:51:20.792403 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" event={"ID":"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d","Type":"ContainerStarted","Data":"44f073bd4c494d5dc1b89f01134e4991f4c5f4d8f89c96dcd9c888bcdb75794d"} Dec 03 06:51:21 crc kubenswrapper[4818]: I1203 06:51:21.357239 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x5c96" Dec 03 06:51:21 crc kubenswrapper[4818]: I1203 06:51:21.449182 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x5c96" Dec 03 06:51:21 crc kubenswrapper[4818]: I1203 06:51:21.814484 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" event={"ID":"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d","Type":"ContainerStarted","Data":"94a0cc2fccdce8f3736a34e3e7aeefb81c8f2a68a5ce8ebb16a6eb0d3067c7be"} Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.176356 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" podStartSLOduration=3.48037633 podStartE2EDuration="4.176333634s" podCreationTimestamp="2025-12-03 06:51:19 +0000 UTC" firstStartedPulling="2025-12-03 06:51:20.745752781 +0000 UTC m=+1438.437361533" lastFinishedPulling="2025-12-03 06:51:21.441710075 +0000 UTC m=+1439.133318837" observedRunningTime="2025-12-03 06:51:21.833618579 +0000 UTC m=+1439.525227341" watchObservedRunningTime="2025-12-03 06:51:23.176333634 +0000 UTC m=+1440.867942406" Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.186027 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x5c96"] Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.187863 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x5c96" podUID="4002e49a-cb3f-477f-bb28-db0ddbc15230" containerName="registry-server" containerID="cri-o://8825dd7dfdbc375d100989a49cccf84cb650673fd46e63d97766601ae8f5a413" gracePeriod=2 Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.627895 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5c96" Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.744774 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4002e49a-cb3f-477f-bb28-db0ddbc15230-utilities\") pod \"4002e49a-cb3f-477f-bb28-db0ddbc15230\" (UID: \"4002e49a-cb3f-477f-bb28-db0ddbc15230\") " Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.745173 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7glsm\" (UniqueName: \"kubernetes.io/projected/4002e49a-cb3f-477f-bb28-db0ddbc15230-kube-api-access-7glsm\") pod \"4002e49a-cb3f-477f-bb28-db0ddbc15230\" (UID: \"4002e49a-cb3f-477f-bb28-db0ddbc15230\") " Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.745254 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4002e49a-cb3f-477f-bb28-db0ddbc15230-catalog-content\") pod \"4002e49a-cb3f-477f-bb28-db0ddbc15230\" (UID: \"4002e49a-cb3f-477f-bb28-db0ddbc15230\") " Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.745452 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4002e49a-cb3f-477f-bb28-db0ddbc15230-utilities" (OuterVolumeSpecName: "utilities") pod "4002e49a-cb3f-477f-bb28-db0ddbc15230" (UID: "4002e49a-cb3f-477f-bb28-db0ddbc15230"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.745983 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4002e49a-cb3f-477f-bb28-db0ddbc15230-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.753579 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4002e49a-cb3f-477f-bb28-db0ddbc15230-kube-api-access-7glsm" (OuterVolumeSpecName: "kube-api-access-7glsm") pod "4002e49a-cb3f-477f-bb28-db0ddbc15230" (UID: "4002e49a-cb3f-477f-bb28-db0ddbc15230"). InnerVolumeSpecName "kube-api-access-7glsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.847352 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7glsm\" (UniqueName: \"kubernetes.io/projected/4002e49a-cb3f-477f-bb28-db0ddbc15230-kube-api-access-7glsm\") on node \"crc\" DevicePath \"\"" Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.852649 4818 generic.go:334] "Generic (PLEG): container finished" podID="4002e49a-cb3f-477f-bb28-db0ddbc15230" containerID="8825dd7dfdbc375d100989a49cccf84cb650673fd46e63d97766601ae8f5a413" exitCode=0 Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.852693 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5c96" event={"ID":"4002e49a-cb3f-477f-bb28-db0ddbc15230","Type":"ContainerDied","Data":"8825dd7dfdbc375d100989a49cccf84cb650673fd46e63d97766601ae8f5a413"} Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.852726 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5c96" event={"ID":"4002e49a-cb3f-477f-bb28-db0ddbc15230","Type":"ContainerDied","Data":"5e626b69cfff97add226a65f3d2bfda160c575ecb03e08d7bf33016ef9e2ea1c"} Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.852749 4818 scope.go:117] "RemoveContainer" containerID="8825dd7dfdbc375d100989a49cccf84cb650673fd46e63d97766601ae8f5a413" Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.853155 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5c96" Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.863519 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4002e49a-cb3f-477f-bb28-db0ddbc15230-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4002e49a-cb3f-477f-bb28-db0ddbc15230" (UID: "4002e49a-cb3f-477f-bb28-db0ddbc15230"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.878739 4818 scope.go:117] "RemoveContainer" containerID="c2fe0f3717047ec1cc4f7793d8718ac0d1977d617ab24837256f54cf7be299a4" Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.911508 4818 scope.go:117] "RemoveContainer" containerID="5615f5b200ddafc1778b890da77f0f90a388c72b7f1d22babcafb5ba6d5a530b" Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.949249 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4002e49a-cb3f-477f-bb28-db0ddbc15230-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.949694 4818 scope.go:117] "RemoveContainer" containerID="8825dd7dfdbc375d100989a49cccf84cb650673fd46e63d97766601ae8f5a413" Dec 03 06:51:23 crc kubenswrapper[4818]: E1203 06:51:23.950219 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8825dd7dfdbc375d100989a49cccf84cb650673fd46e63d97766601ae8f5a413\": container with ID starting with 8825dd7dfdbc375d100989a49cccf84cb650673fd46e63d97766601ae8f5a413 not found: ID does not exist" containerID="8825dd7dfdbc375d100989a49cccf84cb650673fd46e63d97766601ae8f5a413" Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.950296 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8825dd7dfdbc375d100989a49cccf84cb650673fd46e63d97766601ae8f5a413"} err="failed to get container status \"8825dd7dfdbc375d100989a49cccf84cb650673fd46e63d97766601ae8f5a413\": rpc error: code = NotFound desc = could not find container \"8825dd7dfdbc375d100989a49cccf84cb650673fd46e63d97766601ae8f5a413\": container with ID starting with 8825dd7dfdbc375d100989a49cccf84cb650673fd46e63d97766601ae8f5a413 not found: ID does not exist" Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.950360 4818 scope.go:117] "RemoveContainer" containerID="c2fe0f3717047ec1cc4f7793d8718ac0d1977d617ab24837256f54cf7be299a4" Dec 03 06:51:23 crc kubenswrapper[4818]: E1203 06:51:23.950841 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2fe0f3717047ec1cc4f7793d8718ac0d1977d617ab24837256f54cf7be299a4\": container with ID starting with c2fe0f3717047ec1cc4f7793d8718ac0d1977d617ab24837256f54cf7be299a4 not found: ID does not exist" containerID="c2fe0f3717047ec1cc4f7793d8718ac0d1977d617ab24837256f54cf7be299a4" Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.950886 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2fe0f3717047ec1cc4f7793d8718ac0d1977d617ab24837256f54cf7be299a4"} err="failed to get container status \"c2fe0f3717047ec1cc4f7793d8718ac0d1977d617ab24837256f54cf7be299a4\": rpc error: code = NotFound desc = could not find container \"c2fe0f3717047ec1cc4f7793d8718ac0d1977d617ab24837256f54cf7be299a4\": container with ID starting with c2fe0f3717047ec1cc4f7793d8718ac0d1977d617ab24837256f54cf7be299a4 not found: ID does not exist" Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.950913 4818 scope.go:117] "RemoveContainer" containerID="5615f5b200ddafc1778b890da77f0f90a388c72b7f1d22babcafb5ba6d5a530b" Dec 03 06:51:23 crc kubenswrapper[4818]: E1203 06:51:23.951258 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5615f5b200ddafc1778b890da77f0f90a388c72b7f1d22babcafb5ba6d5a530b\": container with ID starting with 5615f5b200ddafc1778b890da77f0f90a388c72b7f1d22babcafb5ba6d5a530b not found: ID does not exist" containerID="5615f5b200ddafc1778b890da77f0f90a388c72b7f1d22babcafb5ba6d5a530b" Dec 03 06:51:23 crc kubenswrapper[4818]: I1203 06:51:23.951286 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5615f5b200ddafc1778b890da77f0f90a388c72b7f1d22babcafb5ba6d5a530b"} err="failed to get container status \"5615f5b200ddafc1778b890da77f0f90a388c72b7f1d22babcafb5ba6d5a530b\": rpc error: code = NotFound desc = could not find container \"5615f5b200ddafc1778b890da77f0f90a388c72b7f1d22babcafb5ba6d5a530b\": container with ID starting with 5615f5b200ddafc1778b890da77f0f90a388c72b7f1d22babcafb5ba6d5a530b not found: ID does not exist" Dec 03 06:51:24 crc kubenswrapper[4818]: I1203 06:51:24.190717 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x5c96"] Dec 03 06:51:24 crc kubenswrapper[4818]: I1203 06:51:24.200869 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x5c96"] Dec 03 06:51:24 crc kubenswrapper[4818]: I1203 06:51:24.751052 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4002e49a-cb3f-477f-bb28-db0ddbc15230" path="/var/lib/kubelet/pods/4002e49a-cb3f-477f-bb28-db0ddbc15230/volumes" Dec 03 06:51:24 crc kubenswrapper[4818]: I1203 06:51:24.863523 4818 generic.go:334] "Generic (PLEG): container finished" podID="fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d" containerID="94a0cc2fccdce8f3736a34e3e7aeefb81c8f2a68a5ce8ebb16a6eb0d3067c7be" exitCode=0 Dec 03 06:51:24 crc kubenswrapper[4818]: I1203 06:51:24.863570 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" event={"ID":"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d","Type":"ContainerDied","Data":"94a0cc2fccdce8f3736a34e3e7aeefb81c8f2a68a5ce8ebb16a6eb0d3067c7be"} Dec 03 06:51:26 crc kubenswrapper[4818]: I1203 06:51:26.351700 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" Dec 03 06:51:26 crc kubenswrapper[4818]: I1203 06:51:26.506185 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-inventory\") pod \"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d\" (UID: \"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d\") " Dec 03 06:51:26 crc kubenswrapper[4818]: I1203 06:51:26.506243 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-ssh-key\") pod \"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d\" (UID: \"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d\") " Dec 03 06:51:26 crc kubenswrapper[4818]: I1203 06:51:26.506437 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnj96\" (UniqueName: \"kubernetes.io/projected/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-kube-api-access-dnj96\") pod \"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d\" (UID: \"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d\") " Dec 03 06:51:26 crc kubenswrapper[4818]: I1203 06:51:26.515061 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-kube-api-access-dnj96" (OuterVolumeSpecName: "kube-api-access-dnj96") pod "fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d" (UID: "fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d"). InnerVolumeSpecName "kube-api-access-dnj96". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:51:26 crc kubenswrapper[4818]: I1203 06:51:26.540457 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-inventory" (OuterVolumeSpecName: "inventory") pod "fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d" (UID: "fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:51:26 crc kubenswrapper[4818]: I1203 06:51:26.545963 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d" (UID: "fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:51:26 crc kubenswrapper[4818]: I1203 06:51:26.626686 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnj96\" (UniqueName: \"kubernetes.io/projected/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-kube-api-access-dnj96\") on node \"crc\" DevicePath \"\"" Dec 03 06:51:26 crc kubenswrapper[4818]: I1203 06:51:26.626734 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 06:51:26 crc kubenswrapper[4818]: I1203 06:51:26.626752 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 06:51:26 crc kubenswrapper[4818]: I1203 06:51:26.887245 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" event={"ID":"fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d","Type":"ContainerDied","Data":"44f073bd4c494d5dc1b89f01134e4991f4c5f4d8f89c96dcd9c888bcdb75794d"} Dec 03 06:51:26 crc kubenswrapper[4818]: I1203 06:51:26.887284 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44f073bd4c494d5dc1b89f01134e4991f4c5f4d8f89c96dcd9c888bcdb75794d" Dec 03 06:51:26 crc kubenswrapper[4818]: I1203 06:51:26.887353 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-nsd6q" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.071292 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g"] Dec 03 06:51:27 crc kubenswrapper[4818]: E1203 06:51:27.071658 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.071674 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 06:51:27 crc kubenswrapper[4818]: E1203 06:51:27.071688 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4002e49a-cb3f-477f-bb28-db0ddbc15230" containerName="extract-content" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.071694 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4002e49a-cb3f-477f-bb28-db0ddbc15230" containerName="extract-content" Dec 03 06:51:27 crc kubenswrapper[4818]: E1203 06:51:27.071706 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4002e49a-cb3f-477f-bb28-db0ddbc15230" containerName="extract-utilities" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.071712 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4002e49a-cb3f-477f-bb28-db0ddbc15230" containerName="extract-utilities" Dec 03 06:51:27 crc kubenswrapper[4818]: E1203 06:51:27.071728 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4002e49a-cb3f-477f-bb28-db0ddbc15230" containerName="registry-server" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.071734 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4002e49a-cb3f-477f-bb28-db0ddbc15230" containerName="registry-server" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.071922 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4002e49a-cb3f-477f-bb28-db0ddbc15230" containerName="registry-server" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.071940 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.073292 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.076049 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.077233 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.077277 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.079599 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2bpsv" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.090232 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g"] Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.172497 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g\" (UID: \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.172668 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdthj\" (UniqueName: \"kubernetes.io/projected/2b2dc637-0ece-412a-b1e2-413dff6a4f08-kube-api-access-rdthj\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g\" (UID: \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.172729 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g\" (UID: \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.172798 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g\" (UID: \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.273773 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdthj\" (UniqueName: \"kubernetes.io/projected/2b2dc637-0ece-412a-b1e2-413dff6a4f08-kube-api-access-rdthj\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g\" (UID: \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.274139 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g\" (UID: \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.274218 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g\" (UID: \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.274319 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g\" (UID: \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.278792 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g\" (UID: \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.279021 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g\" (UID: \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.280297 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g\" (UID: \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.297249 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdthj\" (UniqueName: \"kubernetes.io/projected/2b2dc637-0ece-412a-b1e2-413dff6a4f08-kube-api-access-rdthj\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g\" (UID: \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.390740 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" Dec 03 06:51:27 crc kubenswrapper[4818]: I1203 06:51:27.948444 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g"] Dec 03 06:51:28 crc kubenswrapper[4818]: I1203 06:51:28.917117 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" event={"ID":"2b2dc637-0ece-412a-b1e2-413dff6a4f08","Type":"ContainerStarted","Data":"36cbd242eabd71e9df7568174a681fcd0c8555f897a024e9a3150898193611a3"} Dec 03 06:51:28 crc kubenswrapper[4818]: I1203 06:51:28.917500 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" event={"ID":"2b2dc637-0ece-412a-b1e2-413dff6a4f08","Type":"ContainerStarted","Data":"3361ea115bbc21444b2c2b1e03bfc30a0b6d48d3842ef2caa4565a284cb016c9"} Dec 03 06:51:53 crc kubenswrapper[4818]: I1203 06:51:53.340804 4818 scope.go:117] "RemoveContainer" containerID="3ac3909bb7ac5d22f52702b363772b4b063da74e54668549e4d542916d951527" Dec 03 06:52:13 crc kubenswrapper[4818]: I1203 06:52:13.302903 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:52:13 crc kubenswrapper[4818]: I1203 06:52:13.303536 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:52:43 crc kubenswrapper[4818]: I1203 06:52:43.302914 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:52:43 crc kubenswrapper[4818]: I1203 06:52:43.303559 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:52:53 crc kubenswrapper[4818]: I1203 06:52:53.493405 4818 scope.go:117] "RemoveContainer" containerID="1ced693b4186c9669b7fde7dbc2e7323ca419d09850500ca4d98afd3fb031afe" Dec 03 06:52:53 crc kubenswrapper[4818]: I1203 06:52:53.514769 4818 scope.go:117] "RemoveContainer" containerID="a27026bc5f78f093258fff7de41856c519084974e28212246f72c061552e254f" Dec 03 06:52:53 crc kubenswrapper[4818]: I1203 06:52:53.551735 4818 scope.go:117] "RemoveContainer" containerID="b75008d167e0556d0fd21ae4fa3d86d4b00028e3a7b752c00c1288d30868b6b8" Dec 03 06:53:13 crc kubenswrapper[4818]: I1203 06:53:13.302390 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 06:53:13 crc kubenswrapper[4818]: I1203 06:53:13.302997 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 06:53:13 crc kubenswrapper[4818]: I1203 06:53:13.303057 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 06:53:13 crc kubenswrapper[4818]: I1203 06:53:13.303978 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01"} pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 06:53:13 crc kubenswrapper[4818]: I1203 06:53:13.304036 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" containerID="cri-o://cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" gracePeriod=600 Dec 03 06:53:13 crc kubenswrapper[4818]: E1203 06:53:13.444003 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:53:13 crc kubenswrapper[4818]: I1203 06:53:13.982693 4818 generic.go:334] "Generic (PLEG): container finished" podID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" exitCode=0 Dec 03 06:53:13 crc kubenswrapper[4818]: I1203 06:53:13.982744 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerDied","Data":"cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01"} Dec 03 06:53:13 crc kubenswrapper[4818]: I1203 06:53:13.982777 4818 scope.go:117] "RemoveContainer" containerID="5bf3e0e65704b43280754ba18a3250980330aab9549a78ae7d9381e3d0130a33" Dec 03 06:53:13 crc kubenswrapper[4818]: I1203 06:53:13.985412 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:53:13 crc kubenswrapper[4818]: E1203 06:53:13.985843 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:53:14 crc kubenswrapper[4818]: I1203 06:53:14.005601 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" podStartSLOduration=107.548232238 podStartE2EDuration="1m48.005581047s" podCreationTimestamp="2025-12-03 06:51:26 +0000 UTC" firstStartedPulling="2025-12-03 06:51:27.959023746 +0000 UTC m=+1445.650632498" lastFinishedPulling="2025-12-03 06:51:28.416372515 +0000 UTC m=+1446.107981307" observedRunningTime="2025-12-03 06:51:28.941735396 +0000 UTC m=+1446.633344148" watchObservedRunningTime="2025-12-03 06:53:14.005581047 +0000 UTC m=+1551.697189799" Dec 03 06:53:19 crc kubenswrapper[4818]: I1203 06:53:19.441728 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nsqsp"] Dec 03 06:53:19 crc kubenswrapper[4818]: I1203 06:53:19.444669 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nsqsp" Dec 03 06:53:19 crc kubenswrapper[4818]: I1203 06:53:19.457843 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nsqsp"] Dec 03 06:53:19 crc kubenswrapper[4818]: I1203 06:53:19.630047 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362a5544-f494-4899-b15b-bfe76aa5748e-utilities\") pod \"redhat-marketplace-nsqsp\" (UID: \"362a5544-f494-4899-b15b-bfe76aa5748e\") " pod="openshift-marketplace/redhat-marketplace-nsqsp" Dec 03 06:53:19 crc kubenswrapper[4818]: I1203 06:53:19.630210 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lwrd\" (UniqueName: \"kubernetes.io/projected/362a5544-f494-4899-b15b-bfe76aa5748e-kube-api-access-9lwrd\") pod \"redhat-marketplace-nsqsp\" (UID: \"362a5544-f494-4899-b15b-bfe76aa5748e\") " pod="openshift-marketplace/redhat-marketplace-nsqsp" Dec 03 06:53:19 crc kubenswrapper[4818]: I1203 06:53:19.630503 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362a5544-f494-4899-b15b-bfe76aa5748e-catalog-content\") pod \"redhat-marketplace-nsqsp\" (UID: \"362a5544-f494-4899-b15b-bfe76aa5748e\") " pod="openshift-marketplace/redhat-marketplace-nsqsp" Dec 03 06:53:19 crc kubenswrapper[4818]: I1203 06:53:19.732593 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362a5544-f494-4899-b15b-bfe76aa5748e-catalog-content\") pod \"redhat-marketplace-nsqsp\" (UID: \"362a5544-f494-4899-b15b-bfe76aa5748e\") " pod="openshift-marketplace/redhat-marketplace-nsqsp" Dec 03 06:53:19 crc kubenswrapper[4818]: I1203 06:53:19.732780 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362a5544-f494-4899-b15b-bfe76aa5748e-utilities\") pod \"redhat-marketplace-nsqsp\" (UID: \"362a5544-f494-4899-b15b-bfe76aa5748e\") " pod="openshift-marketplace/redhat-marketplace-nsqsp" Dec 03 06:53:19 crc kubenswrapper[4818]: I1203 06:53:19.732849 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lwrd\" (UniqueName: \"kubernetes.io/projected/362a5544-f494-4899-b15b-bfe76aa5748e-kube-api-access-9lwrd\") pod \"redhat-marketplace-nsqsp\" (UID: \"362a5544-f494-4899-b15b-bfe76aa5748e\") " pod="openshift-marketplace/redhat-marketplace-nsqsp" Dec 03 06:53:19 crc kubenswrapper[4818]: I1203 06:53:19.733259 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362a5544-f494-4899-b15b-bfe76aa5748e-catalog-content\") pod \"redhat-marketplace-nsqsp\" (UID: \"362a5544-f494-4899-b15b-bfe76aa5748e\") " pod="openshift-marketplace/redhat-marketplace-nsqsp" Dec 03 06:53:19 crc kubenswrapper[4818]: I1203 06:53:19.733316 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362a5544-f494-4899-b15b-bfe76aa5748e-utilities\") pod \"redhat-marketplace-nsqsp\" (UID: \"362a5544-f494-4899-b15b-bfe76aa5748e\") " pod="openshift-marketplace/redhat-marketplace-nsqsp" Dec 03 06:53:19 crc kubenswrapper[4818]: I1203 06:53:19.751849 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lwrd\" (UniqueName: \"kubernetes.io/projected/362a5544-f494-4899-b15b-bfe76aa5748e-kube-api-access-9lwrd\") pod \"redhat-marketplace-nsqsp\" (UID: \"362a5544-f494-4899-b15b-bfe76aa5748e\") " pod="openshift-marketplace/redhat-marketplace-nsqsp" Dec 03 06:53:19 crc kubenswrapper[4818]: I1203 06:53:19.825512 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nsqsp" Dec 03 06:53:20 crc kubenswrapper[4818]: I1203 06:53:20.284912 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nsqsp"] Dec 03 06:53:21 crc kubenswrapper[4818]: I1203 06:53:21.057895 4818 generic.go:334] "Generic (PLEG): container finished" podID="362a5544-f494-4899-b15b-bfe76aa5748e" containerID="3409e5a57edbc43330516919506f98580bf071f555bb0de38f79fbbdd873a125" exitCode=0 Dec 03 06:53:21 crc kubenswrapper[4818]: I1203 06:53:21.058000 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nsqsp" event={"ID":"362a5544-f494-4899-b15b-bfe76aa5748e","Type":"ContainerDied","Data":"3409e5a57edbc43330516919506f98580bf071f555bb0de38f79fbbdd873a125"} Dec 03 06:53:21 crc kubenswrapper[4818]: I1203 06:53:21.058289 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nsqsp" event={"ID":"362a5544-f494-4899-b15b-bfe76aa5748e","Type":"ContainerStarted","Data":"6ff93526e478807dd943e9d133056199152e6dcec29cacbb11680cdbb2640174"} Dec 03 06:53:22 crc kubenswrapper[4818]: I1203 06:53:22.069784 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nsqsp" event={"ID":"362a5544-f494-4899-b15b-bfe76aa5748e","Type":"ContainerStarted","Data":"d0d0746fa4052987d093531da874c223056416267326782fcdc4d6264ab1fc0b"} Dec 03 06:53:23 crc kubenswrapper[4818]: I1203 06:53:23.082410 4818 generic.go:334] "Generic (PLEG): container finished" podID="362a5544-f494-4899-b15b-bfe76aa5748e" containerID="d0d0746fa4052987d093531da874c223056416267326782fcdc4d6264ab1fc0b" exitCode=0 Dec 03 06:53:23 crc kubenswrapper[4818]: I1203 06:53:23.082475 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nsqsp" event={"ID":"362a5544-f494-4899-b15b-bfe76aa5748e","Type":"ContainerDied","Data":"d0d0746fa4052987d093531da874c223056416267326782fcdc4d6264ab1fc0b"} Dec 03 06:53:24 crc kubenswrapper[4818]: I1203 06:53:24.097361 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nsqsp" event={"ID":"362a5544-f494-4899-b15b-bfe76aa5748e","Type":"ContainerStarted","Data":"a1eb6e623426bc9fd6980a4eb698300af9920713aea21688223fd11a0c39daca"} Dec 03 06:53:24 crc kubenswrapper[4818]: I1203 06:53:24.121023 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nsqsp" podStartSLOduration=2.653850849 podStartE2EDuration="5.120999296s" podCreationTimestamp="2025-12-03 06:53:19 +0000 UTC" firstStartedPulling="2025-12-03 06:53:21.059912694 +0000 UTC m=+1558.751521446" lastFinishedPulling="2025-12-03 06:53:23.527061131 +0000 UTC m=+1561.218669893" observedRunningTime="2025-12-03 06:53:24.115426839 +0000 UTC m=+1561.807035591" watchObservedRunningTime="2025-12-03 06:53:24.120999296 +0000 UTC m=+1561.812608048" Dec 03 06:53:24 crc kubenswrapper[4818]: I1203 06:53:24.737874 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:53:24 crc kubenswrapper[4818]: E1203 06:53:24.738469 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:53:29 crc kubenswrapper[4818]: I1203 06:53:29.396960 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sx949"] Dec 03 06:53:29 crc kubenswrapper[4818]: I1203 06:53:29.399949 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sx949" Dec 03 06:53:29 crc kubenswrapper[4818]: I1203 06:53:29.419799 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sx949"] Dec 03 06:53:29 crc kubenswrapper[4818]: I1203 06:53:29.538557 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnxdv\" (UniqueName: \"kubernetes.io/projected/da7666f8-66ae-402f-9845-7e0fbebdc8a0-kube-api-access-nnxdv\") pod \"community-operators-sx949\" (UID: \"da7666f8-66ae-402f-9845-7e0fbebdc8a0\") " pod="openshift-marketplace/community-operators-sx949" Dec 03 06:53:29 crc kubenswrapper[4818]: I1203 06:53:29.538684 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da7666f8-66ae-402f-9845-7e0fbebdc8a0-catalog-content\") pod \"community-operators-sx949\" (UID: \"da7666f8-66ae-402f-9845-7e0fbebdc8a0\") " pod="openshift-marketplace/community-operators-sx949" Dec 03 06:53:29 crc kubenswrapper[4818]: I1203 06:53:29.538794 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da7666f8-66ae-402f-9845-7e0fbebdc8a0-utilities\") pod \"community-operators-sx949\" (UID: \"da7666f8-66ae-402f-9845-7e0fbebdc8a0\") " pod="openshift-marketplace/community-operators-sx949" Dec 03 06:53:29 crc kubenswrapper[4818]: I1203 06:53:29.640028 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnxdv\" (UniqueName: \"kubernetes.io/projected/da7666f8-66ae-402f-9845-7e0fbebdc8a0-kube-api-access-nnxdv\") pod \"community-operators-sx949\" (UID: \"da7666f8-66ae-402f-9845-7e0fbebdc8a0\") " pod="openshift-marketplace/community-operators-sx949" Dec 03 06:53:29 crc kubenswrapper[4818]: I1203 06:53:29.640122 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da7666f8-66ae-402f-9845-7e0fbebdc8a0-catalog-content\") pod \"community-operators-sx949\" (UID: \"da7666f8-66ae-402f-9845-7e0fbebdc8a0\") " pod="openshift-marketplace/community-operators-sx949" Dec 03 06:53:29 crc kubenswrapper[4818]: I1203 06:53:29.640201 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da7666f8-66ae-402f-9845-7e0fbebdc8a0-utilities\") pod \"community-operators-sx949\" (UID: \"da7666f8-66ae-402f-9845-7e0fbebdc8a0\") " pod="openshift-marketplace/community-operators-sx949" Dec 03 06:53:29 crc kubenswrapper[4818]: I1203 06:53:29.640693 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da7666f8-66ae-402f-9845-7e0fbebdc8a0-utilities\") pod \"community-operators-sx949\" (UID: \"da7666f8-66ae-402f-9845-7e0fbebdc8a0\") " pod="openshift-marketplace/community-operators-sx949" Dec 03 06:53:29 crc kubenswrapper[4818]: I1203 06:53:29.640933 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da7666f8-66ae-402f-9845-7e0fbebdc8a0-catalog-content\") pod \"community-operators-sx949\" (UID: \"da7666f8-66ae-402f-9845-7e0fbebdc8a0\") " pod="openshift-marketplace/community-operators-sx949" Dec 03 06:53:29 crc kubenswrapper[4818]: I1203 06:53:29.665602 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnxdv\" (UniqueName: \"kubernetes.io/projected/da7666f8-66ae-402f-9845-7e0fbebdc8a0-kube-api-access-nnxdv\") pod \"community-operators-sx949\" (UID: \"da7666f8-66ae-402f-9845-7e0fbebdc8a0\") " pod="openshift-marketplace/community-operators-sx949" Dec 03 06:53:29 crc kubenswrapper[4818]: I1203 06:53:29.720427 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sx949" Dec 03 06:53:29 crc kubenswrapper[4818]: I1203 06:53:29.825619 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nsqsp" Dec 03 06:53:29 crc kubenswrapper[4818]: I1203 06:53:29.826010 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nsqsp" Dec 03 06:53:29 crc kubenswrapper[4818]: I1203 06:53:29.915618 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nsqsp" Dec 03 06:53:30 crc kubenswrapper[4818]: I1203 06:53:30.250407 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nsqsp" Dec 03 06:53:30 crc kubenswrapper[4818]: I1203 06:53:30.270117 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sx949"] Dec 03 06:53:31 crc kubenswrapper[4818]: I1203 06:53:31.202877 4818 generic.go:334] "Generic (PLEG): container finished" podID="da7666f8-66ae-402f-9845-7e0fbebdc8a0" containerID="a68aa1c131b00b50aac1da263bbbefb1daffc3752130ab8d0e30e49a5a393808" exitCode=0 Dec 03 06:53:31 crc kubenswrapper[4818]: I1203 06:53:31.202941 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sx949" event={"ID":"da7666f8-66ae-402f-9845-7e0fbebdc8a0","Type":"ContainerDied","Data":"a68aa1c131b00b50aac1da263bbbefb1daffc3752130ab8d0e30e49a5a393808"} Dec 03 06:53:31 crc kubenswrapper[4818]: I1203 06:53:31.203200 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sx949" event={"ID":"da7666f8-66ae-402f-9845-7e0fbebdc8a0","Type":"ContainerStarted","Data":"bf5f93b06d1e6afab41ee9a8708f51831f5ce4f350057f76b4d66fe06acd8569"} Dec 03 06:53:32 crc kubenswrapper[4818]: I1203 06:53:32.180449 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nsqsp"] Dec 03 06:53:32 crc kubenswrapper[4818]: I1203 06:53:32.215063 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sx949" event={"ID":"da7666f8-66ae-402f-9845-7e0fbebdc8a0","Type":"ContainerStarted","Data":"63a30d131f038d6360c3e6683c6a638f0c310f6b9e2deec36f20fac75e4eac3f"} Dec 03 06:53:32 crc kubenswrapper[4818]: I1203 06:53:32.215295 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nsqsp" podUID="362a5544-f494-4899-b15b-bfe76aa5748e" containerName="registry-server" containerID="cri-o://a1eb6e623426bc9fd6980a4eb698300af9920713aea21688223fd11a0c39daca" gracePeriod=2 Dec 03 06:53:32 crc kubenswrapper[4818]: I1203 06:53:32.669214 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nsqsp" Dec 03 06:53:32 crc kubenswrapper[4818]: I1203 06:53:32.753823 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362a5544-f494-4899-b15b-bfe76aa5748e-utilities\") pod \"362a5544-f494-4899-b15b-bfe76aa5748e\" (UID: \"362a5544-f494-4899-b15b-bfe76aa5748e\") " Dec 03 06:53:32 crc kubenswrapper[4818]: I1203 06:53:32.753939 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lwrd\" (UniqueName: \"kubernetes.io/projected/362a5544-f494-4899-b15b-bfe76aa5748e-kube-api-access-9lwrd\") pod \"362a5544-f494-4899-b15b-bfe76aa5748e\" (UID: \"362a5544-f494-4899-b15b-bfe76aa5748e\") " Dec 03 06:53:32 crc kubenswrapper[4818]: I1203 06:53:32.753978 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362a5544-f494-4899-b15b-bfe76aa5748e-catalog-content\") pod \"362a5544-f494-4899-b15b-bfe76aa5748e\" (UID: \"362a5544-f494-4899-b15b-bfe76aa5748e\") " Dec 03 06:53:32 crc kubenswrapper[4818]: I1203 06:53:32.754925 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/362a5544-f494-4899-b15b-bfe76aa5748e-utilities" (OuterVolumeSpecName: "utilities") pod "362a5544-f494-4899-b15b-bfe76aa5748e" (UID: "362a5544-f494-4899-b15b-bfe76aa5748e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:53:32 crc kubenswrapper[4818]: I1203 06:53:32.760693 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/362a5544-f494-4899-b15b-bfe76aa5748e-kube-api-access-9lwrd" (OuterVolumeSpecName: "kube-api-access-9lwrd") pod "362a5544-f494-4899-b15b-bfe76aa5748e" (UID: "362a5544-f494-4899-b15b-bfe76aa5748e"). InnerVolumeSpecName "kube-api-access-9lwrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:53:32 crc kubenswrapper[4818]: I1203 06:53:32.775468 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/362a5544-f494-4899-b15b-bfe76aa5748e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "362a5544-f494-4899-b15b-bfe76aa5748e" (UID: "362a5544-f494-4899-b15b-bfe76aa5748e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:53:32 crc kubenswrapper[4818]: I1203 06:53:32.856166 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362a5544-f494-4899-b15b-bfe76aa5748e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:32 crc kubenswrapper[4818]: I1203 06:53:32.856216 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lwrd\" (UniqueName: \"kubernetes.io/projected/362a5544-f494-4899-b15b-bfe76aa5748e-kube-api-access-9lwrd\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:32 crc kubenswrapper[4818]: I1203 06:53:32.856241 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362a5544-f494-4899-b15b-bfe76aa5748e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:33 crc kubenswrapper[4818]: I1203 06:53:33.227625 4818 generic.go:334] "Generic (PLEG): container finished" podID="362a5544-f494-4899-b15b-bfe76aa5748e" containerID="a1eb6e623426bc9fd6980a4eb698300af9920713aea21688223fd11a0c39daca" exitCode=0 Dec 03 06:53:33 crc kubenswrapper[4818]: I1203 06:53:33.227698 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nsqsp" event={"ID":"362a5544-f494-4899-b15b-bfe76aa5748e","Type":"ContainerDied","Data":"a1eb6e623426bc9fd6980a4eb698300af9920713aea21688223fd11a0c39daca"} Dec 03 06:53:33 crc kubenswrapper[4818]: I1203 06:53:33.227728 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nsqsp" event={"ID":"362a5544-f494-4899-b15b-bfe76aa5748e","Type":"ContainerDied","Data":"6ff93526e478807dd943e9d133056199152e6dcec29cacbb11680cdbb2640174"} Dec 03 06:53:33 crc kubenswrapper[4818]: I1203 06:53:33.227748 4818 scope.go:117] "RemoveContainer" containerID="a1eb6e623426bc9fd6980a4eb698300af9920713aea21688223fd11a0c39daca" Dec 03 06:53:33 crc kubenswrapper[4818]: I1203 06:53:33.227760 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nsqsp" Dec 03 06:53:33 crc kubenswrapper[4818]: I1203 06:53:33.232309 4818 generic.go:334] "Generic (PLEG): container finished" podID="da7666f8-66ae-402f-9845-7e0fbebdc8a0" containerID="63a30d131f038d6360c3e6683c6a638f0c310f6b9e2deec36f20fac75e4eac3f" exitCode=0 Dec 03 06:53:33 crc kubenswrapper[4818]: I1203 06:53:33.232342 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sx949" event={"ID":"da7666f8-66ae-402f-9845-7e0fbebdc8a0","Type":"ContainerDied","Data":"63a30d131f038d6360c3e6683c6a638f0c310f6b9e2deec36f20fac75e4eac3f"} Dec 03 06:53:33 crc kubenswrapper[4818]: I1203 06:53:33.289027 4818 scope.go:117] "RemoveContainer" containerID="d0d0746fa4052987d093531da874c223056416267326782fcdc4d6264ab1fc0b" Dec 03 06:53:33 crc kubenswrapper[4818]: I1203 06:53:33.294394 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nsqsp"] Dec 03 06:53:33 crc kubenswrapper[4818]: I1203 06:53:33.308458 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nsqsp"] Dec 03 06:53:33 crc kubenswrapper[4818]: I1203 06:53:33.311482 4818 scope.go:117] "RemoveContainer" containerID="3409e5a57edbc43330516919506f98580bf071f555bb0de38f79fbbdd873a125" Dec 03 06:53:33 crc kubenswrapper[4818]: I1203 06:53:33.359584 4818 scope.go:117] "RemoveContainer" containerID="a1eb6e623426bc9fd6980a4eb698300af9920713aea21688223fd11a0c39daca" Dec 03 06:53:33 crc kubenswrapper[4818]: E1203 06:53:33.360059 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1eb6e623426bc9fd6980a4eb698300af9920713aea21688223fd11a0c39daca\": container with ID starting with a1eb6e623426bc9fd6980a4eb698300af9920713aea21688223fd11a0c39daca not found: ID does not exist" containerID="a1eb6e623426bc9fd6980a4eb698300af9920713aea21688223fd11a0c39daca" Dec 03 06:53:33 crc kubenswrapper[4818]: I1203 06:53:33.360117 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1eb6e623426bc9fd6980a4eb698300af9920713aea21688223fd11a0c39daca"} err="failed to get container status \"a1eb6e623426bc9fd6980a4eb698300af9920713aea21688223fd11a0c39daca\": rpc error: code = NotFound desc = could not find container \"a1eb6e623426bc9fd6980a4eb698300af9920713aea21688223fd11a0c39daca\": container with ID starting with a1eb6e623426bc9fd6980a4eb698300af9920713aea21688223fd11a0c39daca not found: ID does not exist" Dec 03 06:53:33 crc kubenswrapper[4818]: I1203 06:53:33.360160 4818 scope.go:117] "RemoveContainer" containerID="d0d0746fa4052987d093531da874c223056416267326782fcdc4d6264ab1fc0b" Dec 03 06:53:33 crc kubenswrapper[4818]: E1203 06:53:33.360639 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0d0746fa4052987d093531da874c223056416267326782fcdc4d6264ab1fc0b\": container with ID starting with d0d0746fa4052987d093531da874c223056416267326782fcdc4d6264ab1fc0b not found: ID does not exist" containerID="d0d0746fa4052987d093531da874c223056416267326782fcdc4d6264ab1fc0b" Dec 03 06:53:33 crc kubenswrapper[4818]: I1203 06:53:33.360675 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0d0746fa4052987d093531da874c223056416267326782fcdc4d6264ab1fc0b"} err="failed to get container status \"d0d0746fa4052987d093531da874c223056416267326782fcdc4d6264ab1fc0b\": rpc error: code = NotFound desc = could not find container \"d0d0746fa4052987d093531da874c223056416267326782fcdc4d6264ab1fc0b\": container with ID starting with d0d0746fa4052987d093531da874c223056416267326782fcdc4d6264ab1fc0b not found: ID does not exist" Dec 03 06:53:33 crc kubenswrapper[4818]: I1203 06:53:33.360702 4818 scope.go:117] "RemoveContainer" containerID="3409e5a57edbc43330516919506f98580bf071f555bb0de38f79fbbdd873a125" Dec 03 06:53:33 crc kubenswrapper[4818]: E1203 06:53:33.361571 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3409e5a57edbc43330516919506f98580bf071f555bb0de38f79fbbdd873a125\": container with ID starting with 3409e5a57edbc43330516919506f98580bf071f555bb0de38f79fbbdd873a125 not found: ID does not exist" containerID="3409e5a57edbc43330516919506f98580bf071f555bb0de38f79fbbdd873a125" Dec 03 06:53:33 crc kubenswrapper[4818]: I1203 06:53:33.361602 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3409e5a57edbc43330516919506f98580bf071f555bb0de38f79fbbdd873a125"} err="failed to get container status \"3409e5a57edbc43330516919506f98580bf071f555bb0de38f79fbbdd873a125\": rpc error: code = NotFound desc = could not find container \"3409e5a57edbc43330516919506f98580bf071f555bb0de38f79fbbdd873a125\": container with ID starting with 3409e5a57edbc43330516919506f98580bf071f555bb0de38f79fbbdd873a125 not found: ID does not exist" Dec 03 06:53:34 crc kubenswrapper[4818]: I1203 06:53:34.245753 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sx949" event={"ID":"da7666f8-66ae-402f-9845-7e0fbebdc8a0","Type":"ContainerStarted","Data":"57decf300a5041acbd784fbfe237560465851bf53ae5f98940a1b9a0d09d9a1f"} Dec 03 06:53:34 crc kubenswrapper[4818]: I1203 06:53:34.288669 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sx949" podStartSLOduration=2.657196859 podStartE2EDuration="5.288637186s" podCreationTimestamp="2025-12-03 06:53:29 +0000 UTC" firstStartedPulling="2025-12-03 06:53:31.204494834 +0000 UTC m=+1568.896103596" lastFinishedPulling="2025-12-03 06:53:33.835935161 +0000 UTC m=+1571.527543923" observedRunningTime="2025-12-03 06:53:34.276462085 +0000 UTC m=+1571.968070857" watchObservedRunningTime="2025-12-03 06:53:34.288637186 +0000 UTC m=+1571.980245978" Dec 03 06:53:34 crc kubenswrapper[4818]: I1203 06:53:34.748804 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="362a5544-f494-4899-b15b-bfe76aa5748e" path="/var/lib/kubelet/pods/362a5544-f494-4899-b15b-bfe76aa5748e/volumes" Dec 03 06:53:36 crc kubenswrapper[4818]: I1203 06:53:36.738674 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:53:36 crc kubenswrapper[4818]: E1203 06:53:36.739568 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:53:39 crc kubenswrapper[4818]: I1203 06:53:39.721738 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sx949" Dec 03 06:53:39 crc kubenswrapper[4818]: I1203 06:53:39.722056 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sx949" Dec 03 06:53:39 crc kubenswrapper[4818]: I1203 06:53:39.799397 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sx949" Dec 03 06:53:40 crc kubenswrapper[4818]: I1203 06:53:40.372752 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sx949" Dec 03 06:53:40 crc kubenswrapper[4818]: I1203 06:53:40.442030 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sx949"] Dec 03 06:53:42 crc kubenswrapper[4818]: I1203 06:53:42.342100 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sx949" podUID="da7666f8-66ae-402f-9845-7e0fbebdc8a0" containerName="registry-server" containerID="cri-o://57decf300a5041acbd784fbfe237560465851bf53ae5f98940a1b9a0d09d9a1f" gracePeriod=2 Dec 03 06:53:43 crc kubenswrapper[4818]: I1203 06:53:43.359671 4818 generic.go:334] "Generic (PLEG): container finished" podID="da7666f8-66ae-402f-9845-7e0fbebdc8a0" containerID="57decf300a5041acbd784fbfe237560465851bf53ae5f98940a1b9a0d09d9a1f" exitCode=0 Dec 03 06:53:43 crc kubenswrapper[4818]: I1203 06:53:43.360038 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sx949" event={"ID":"da7666f8-66ae-402f-9845-7e0fbebdc8a0","Type":"ContainerDied","Data":"57decf300a5041acbd784fbfe237560465851bf53ae5f98940a1b9a0d09d9a1f"} Dec 03 06:53:43 crc kubenswrapper[4818]: I1203 06:53:43.418070 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sx949" Dec 03 06:53:43 crc kubenswrapper[4818]: I1203 06:53:43.618703 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da7666f8-66ae-402f-9845-7e0fbebdc8a0-catalog-content\") pod \"da7666f8-66ae-402f-9845-7e0fbebdc8a0\" (UID: \"da7666f8-66ae-402f-9845-7e0fbebdc8a0\") " Dec 03 06:53:43 crc kubenswrapper[4818]: I1203 06:53:43.619162 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da7666f8-66ae-402f-9845-7e0fbebdc8a0-utilities\") pod \"da7666f8-66ae-402f-9845-7e0fbebdc8a0\" (UID: \"da7666f8-66ae-402f-9845-7e0fbebdc8a0\") " Dec 03 06:53:43 crc kubenswrapper[4818]: I1203 06:53:43.620129 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da7666f8-66ae-402f-9845-7e0fbebdc8a0-utilities" (OuterVolumeSpecName: "utilities") pod "da7666f8-66ae-402f-9845-7e0fbebdc8a0" (UID: "da7666f8-66ae-402f-9845-7e0fbebdc8a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:53:43 crc kubenswrapper[4818]: I1203 06:53:43.620195 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnxdv\" (UniqueName: \"kubernetes.io/projected/da7666f8-66ae-402f-9845-7e0fbebdc8a0-kube-api-access-nnxdv\") pod \"da7666f8-66ae-402f-9845-7e0fbebdc8a0\" (UID: \"da7666f8-66ae-402f-9845-7e0fbebdc8a0\") " Dec 03 06:53:43 crc kubenswrapper[4818]: I1203 06:53:43.621493 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da7666f8-66ae-402f-9845-7e0fbebdc8a0-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:43 crc kubenswrapper[4818]: I1203 06:53:43.627198 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da7666f8-66ae-402f-9845-7e0fbebdc8a0-kube-api-access-nnxdv" (OuterVolumeSpecName: "kube-api-access-nnxdv") pod "da7666f8-66ae-402f-9845-7e0fbebdc8a0" (UID: "da7666f8-66ae-402f-9845-7e0fbebdc8a0"). InnerVolumeSpecName "kube-api-access-nnxdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:53:43 crc kubenswrapper[4818]: I1203 06:53:43.675775 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da7666f8-66ae-402f-9845-7e0fbebdc8a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da7666f8-66ae-402f-9845-7e0fbebdc8a0" (UID: "da7666f8-66ae-402f-9845-7e0fbebdc8a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:53:43 crc kubenswrapper[4818]: I1203 06:53:43.723039 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnxdv\" (UniqueName: \"kubernetes.io/projected/da7666f8-66ae-402f-9845-7e0fbebdc8a0-kube-api-access-nnxdv\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:43 crc kubenswrapper[4818]: I1203 06:53:43.723091 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da7666f8-66ae-402f-9845-7e0fbebdc8a0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:53:44 crc kubenswrapper[4818]: I1203 06:53:44.379765 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sx949" event={"ID":"da7666f8-66ae-402f-9845-7e0fbebdc8a0","Type":"ContainerDied","Data":"bf5f93b06d1e6afab41ee9a8708f51831f5ce4f350057f76b4d66fe06acd8569"} Dec 03 06:53:44 crc kubenswrapper[4818]: I1203 06:53:44.379856 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sx949" Dec 03 06:53:44 crc kubenswrapper[4818]: I1203 06:53:44.379892 4818 scope.go:117] "RemoveContainer" containerID="57decf300a5041acbd784fbfe237560465851bf53ae5f98940a1b9a0d09d9a1f" Dec 03 06:53:44 crc kubenswrapper[4818]: I1203 06:53:44.403148 4818 scope.go:117] "RemoveContainer" containerID="63a30d131f038d6360c3e6683c6a638f0c310f6b9e2deec36f20fac75e4eac3f" Dec 03 06:53:44 crc kubenswrapper[4818]: I1203 06:53:44.430807 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sx949"] Dec 03 06:53:44 crc kubenswrapper[4818]: I1203 06:53:44.445843 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sx949"] Dec 03 06:53:44 crc kubenswrapper[4818]: I1203 06:53:44.460926 4818 scope.go:117] "RemoveContainer" containerID="a68aa1c131b00b50aac1da263bbbefb1daffc3752130ab8d0e30e49a5a393808" Dec 03 06:53:44 crc kubenswrapper[4818]: I1203 06:53:44.753697 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da7666f8-66ae-402f-9845-7e0fbebdc8a0" path="/var/lib/kubelet/pods/da7666f8-66ae-402f-9845-7e0fbebdc8a0/volumes" Dec 03 06:53:50 crc kubenswrapper[4818]: I1203 06:53:50.738200 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:53:50 crc kubenswrapper[4818]: E1203 06:53:50.739044 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:54:02 crc kubenswrapper[4818]: I1203 06:54:02.743801 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:54:02 crc kubenswrapper[4818]: E1203 06:54:02.744637 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:54:17 crc kubenswrapper[4818]: I1203 06:54:17.738383 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:54:17 crc kubenswrapper[4818]: E1203 06:54:17.739282 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:54:30 crc kubenswrapper[4818]: I1203 06:54:30.738008 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:54:30 crc kubenswrapper[4818]: E1203 06:54:30.738877 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:54:42 crc kubenswrapper[4818]: I1203 06:54:42.744834 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:54:42 crc kubenswrapper[4818]: E1203 06:54:42.745805 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.155075 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4ndx9"] Dec 03 06:54:46 crc kubenswrapper[4818]: E1203 06:54:46.156024 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="362a5544-f494-4899-b15b-bfe76aa5748e" containerName="extract-utilities" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.156039 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="362a5544-f494-4899-b15b-bfe76aa5748e" containerName="extract-utilities" Dec 03 06:54:46 crc kubenswrapper[4818]: E1203 06:54:46.156056 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da7666f8-66ae-402f-9845-7e0fbebdc8a0" containerName="extract-content" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.156062 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="da7666f8-66ae-402f-9845-7e0fbebdc8a0" containerName="extract-content" Dec 03 06:54:46 crc kubenswrapper[4818]: E1203 06:54:46.156073 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da7666f8-66ae-402f-9845-7e0fbebdc8a0" containerName="registry-server" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.156080 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="da7666f8-66ae-402f-9845-7e0fbebdc8a0" containerName="registry-server" Dec 03 06:54:46 crc kubenswrapper[4818]: E1203 06:54:46.156103 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="362a5544-f494-4899-b15b-bfe76aa5748e" containerName="extract-content" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.156114 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="362a5544-f494-4899-b15b-bfe76aa5748e" containerName="extract-content" Dec 03 06:54:46 crc kubenswrapper[4818]: E1203 06:54:46.156140 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="362a5544-f494-4899-b15b-bfe76aa5748e" containerName="registry-server" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.156146 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="362a5544-f494-4899-b15b-bfe76aa5748e" containerName="registry-server" Dec 03 06:54:46 crc kubenswrapper[4818]: E1203 06:54:46.156161 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da7666f8-66ae-402f-9845-7e0fbebdc8a0" containerName="extract-utilities" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.156167 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="da7666f8-66ae-402f-9845-7e0fbebdc8a0" containerName="extract-utilities" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.156375 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="362a5544-f494-4899-b15b-bfe76aa5748e" containerName="registry-server" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.156402 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="da7666f8-66ae-402f-9845-7e0fbebdc8a0" containerName="registry-server" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.157960 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4ndx9" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.170241 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4ndx9"] Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.241177 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6276f22e-c3ff-4d6c-a1b2-02119f513101-utilities\") pod \"certified-operators-4ndx9\" (UID: \"6276f22e-c3ff-4d6c-a1b2-02119f513101\") " pod="openshift-marketplace/certified-operators-4ndx9" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.241248 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6276f22e-c3ff-4d6c-a1b2-02119f513101-catalog-content\") pod \"certified-operators-4ndx9\" (UID: \"6276f22e-c3ff-4d6c-a1b2-02119f513101\") " pod="openshift-marketplace/certified-operators-4ndx9" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.241390 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5hpc\" (UniqueName: \"kubernetes.io/projected/6276f22e-c3ff-4d6c-a1b2-02119f513101-kube-api-access-k5hpc\") pod \"certified-operators-4ndx9\" (UID: \"6276f22e-c3ff-4d6c-a1b2-02119f513101\") " pod="openshift-marketplace/certified-operators-4ndx9" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.343128 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6276f22e-c3ff-4d6c-a1b2-02119f513101-utilities\") pod \"certified-operators-4ndx9\" (UID: \"6276f22e-c3ff-4d6c-a1b2-02119f513101\") " pod="openshift-marketplace/certified-operators-4ndx9" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.343224 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6276f22e-c3ff-4d6c-a1b2-02119f513101-catalog-content\") pod \"certified-operators-4ndx9\" (UID: \"6276f22e-c3ff-4d6c-a1b2-02119f513101\") " pod="openshift-marketplace/certified-operators-4ndx9" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.343321 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5hpc\" (UniqueName: \"kubernetes.io/projected/6276f22e-c3ff-4d6c-a1b2-02119f513101-kube-api-access-k5hpc\") pod \"certified-operators-4ndx9\" (UID: \"6276f22e-c3ff-4d6c-a1b2-02119f513101\") " pod="openshift-marketplace/certified-operators-4ndx9" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.343899 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6276f22e-c3ff-4d6c-a1b2-02119f513101-catalog-content\") pod \"certified-operators-4ndx9\" (UID: \"6276f22e-c3ff-4d6c-a1b2-02119f513101\") " pod="openshift-marketplace/certified-operators-4ndx9" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.344097 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6276f22e-c3ff-4d6c-a1b2-02119f513101-utilities\") pod \"certified-operators-4ndx9\" (UID: \"6276f22e-c3ff-4d6c-a1b2-02119f513101\") " pod="openshift-marketplace/certified-operators-4ndx9" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.373974 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5hpc\" (UniqueName: \"kubernetes.io/projected/6276f22e-c3ff-4d6c-a1b2-02119f513101-kube-api-access-k5hpc\") pod \"certified-operators-4ndx9\" (UID: \"6276f22e-c3ff-4d6c-a1b2-02119f513101\") " pod="openshift-marketplace/certified-operators-4ndx9" Dec 03 06:54:46 crc kubenswrapper[4818]: I1203 06:54:46.489543 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4ndx9" Dec 03 06:54:47 crc kubenswrapper[4818]: I1203 06:54:47.013437 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4ndx9"] Dec 03 06:54:47 crc kubenswrapper[4818]: I1203 06:54:47.173777 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ndx9" event={"ID":"6276f22e-c3ff-4d6c-a1b2-02119f513101","Type":"ContainerStarted","Data":"65817d01efafdcc1ee47ea9d9f9358f503d41725b960dae4d34f95a28b2de8f4"} Dec 03 06:54:48 crc kubenswrapper[4818]: I1203 06:54:48.189401 4818 generic.go:334] "Generic (PLEG): container finished" podID="2b2dc637-0ece-412a-b1e2-413dff6a4f08" containerID="36cbd242eabd71e9df7568174a681fcd0c8555f897a024e9a3150898193611a3" exitCode=0 Dec 03 06:54:48 crc kubenswrapper[4818]: I1203 06:54:48.189486 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" event={"ID":"2b2dc637-0ece-412a-b1e2-413dff6a4f08","Type":"ContainerDied","Data":"36cbd242eabd71e9df7568174a681fcd0c8555f897a024e9a3150898193611a3"} Dec 03 06:54:48 crc kubenswrapper[4818]: I1203 06:54:48.192609 4818 generic.go:334] "Generic (PLEG): container finished" podID="6276f22e-c3ff-4d6c-a1b2-02119f513101" containerID="4bc7487abd396c9bb7c25deb1830a45c6d7357828f002f2b410da52f563b75a6" exitCode=0 Dec 03 06:54:48 crc kubenswrapper[4818]: I1203 06:54:48.192674 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ndx9" event={"ID":"6276f22e-c3ff-4d6c-a1b2-02119f513101","Type":"ContainerDied","Data":"4bc7487abd396c9bb7c25deb1830a45c6d7357828f002f2b410da52f563b75a6"} Dec 03 06:54:48 crc kubenswrapper[4818]: I1203 06:54:48.195763 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 06:54:49 crc kubenswrapper[4818]: I1203 06:54:49.206574 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ndx9" event={"ID":"6276f22e-c3ff-4d6c-a1b2-02119f513101","Type":"ContainerStarted","Data":"2a1c3436935aa1aad8fcd2df436373a98c3a928a0ab3ac6dc765312838671b51"} Dec 03 06:54:49 crc kubenswrapper[4818]: I1203 06:54:49.712107 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" Dec 03 06:54:49 crc kubenswrapper[4818]: I1203 06:54:49.744035 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-bootstrap-combined-ca-bundle\") pod \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\" (UID: \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\") " Dec 03 06:54:49 crc kubenswrapper[4818]: I1203 06:54:49.744217 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdthj\" (UniqueName: \"kubernetes.io/projected/2b2dc637-0ece-412a-b1e2-413dff6a4f08-kube-api-access-rdthj\") pod \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\" (UID: \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\") " Dec 03 06:54:49 crc kubenswrapper[4818]: I1203 06:54:49.744318 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-inventory\") pod \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\" (UID: \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\") " Dec 03 06:54:49 crc kubenswrapper[4818]: I1203 06:54:49.744392 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-ssh-key\") pod \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\" (UID: \"2b2dc637-0ece-412a-b1e2-413dff6a4f08\") " Dec 03 06:54:49 crc kubenswrapper[4818]: I1203 06:54:49.750166 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b2dc637-0ece-412a-b1e2-413dff6a4f08-kube-api-access-rdthj" (OuterVolumeSpecName: "kube-api-access-rdthj") pod "2b2dc637-0ece-412a-b1e2-413dff6a4f08" (UID: "2b2dc637-0ece-412a-b1e2-413dff6a4f08"). InnerVolumeSpecName "kube-api-access-rdthj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:54:49 crc kubenswrapper[4818]: I1203 06:54:49.750884 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "2b2dc637-0ece-412a-b1e2-413dff6a4f08" (UID: "2b2dc637-0ece-412a-b1e2-413dff6a4f08"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:54:49 crc kubenswrapper[4818]: I1203 06:54:49.776573 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-inventory" (OuterVolumeSpecName: "inventory") pod "2b2dc637-0ece-412a-b1e2-413dff6a4f08" (UID: "2b2dc637-0ece-412a-b1e2-413dff6a4f08"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:54:49 crc kubenswrapper[4818]: I1203 06:54:49.778248 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2b2dc637-0ece-412a-b1e2-413dff6a4f08" (UID: "2b2dc637-0ece-412a-b1e2-413dff6a4f08"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:54:49 crc kubenswrapper[4818]: I1203 06:54:49.847475 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 06:54:49 crc kubenswrapper[4818]: I1203 06:54:49.847505 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 06:54:49 crc kubenswrapper[4818]: I1203 06:54:49.847515 4818 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b2dc637-0ece-412a-b1e2-413dff6a4f08-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 06:54:49 crc kubenswrapper[4818]: I1203 06:54:49.847525 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdthj\" (UniqueName: \"kubernetes.io/projected/2b2dc637-0ece-412a-b1e2-413dff6a4f08-kube-api-access-rdthj\") on node \"crc\" DevicePath \"\"" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.221868 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" event={"ID":"2b2dc637-0ece-412a-b1e2-413dff6a4f08","Type":"ContainerDied","Data":"3361ea115bbc21444b2c2b1e03bfc30a0b6d48d3842ef2caa4565a284cb016c9"} Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.221929 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.221931 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3361ea115bbc21444b2c2b1e03bfc30a0b6d48d3842ef2caa4565a284cb016c9" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.225280 4818 generic.go:334] "Generic (PLEG): container finished" podID="6276f22e-c3ff-4d6c-a1b2-02119f513101" containerID="2a1c3436935aa1aad8fcd2df436373a98c3a928a0ab3ac6dc765312838671b51" exitCode=0 Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.225380 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ndx9" event={"ID":"6276f22e-c3ff-4d6c-a1b2-02119f513101","Type":"ContainerDied","Data":"2a1c3436935aa1aad8fcd2df436373a98c3a928a0ab3ac6dc765312838671b51"} Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.336564 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq"] Dec 03 06:54:50 crc kubenswrapper[4818]: E1203 06:54:50.337011 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b2dc637-0ece-412a-b1e2-413dff6a4f08" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.337033 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b2dc637-0ece-412a-b1e2-413dff6a4f08" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.337325 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b2dc637-0ece-412a-b1e2-413dff6a4f08" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.338065 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.341364 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.341912 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.343180 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.344490 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2bpsv" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.357250 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq"] Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.360592 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86dc24e9-831f-4731-a612-98009562639c-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq\" (UID: \"86dc24e9-831f-4731-a612-98009562639c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.360700 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86dc24e9-831f-4731-a612-98009562639c-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq\" (UID: \"86dc24e9-831f-4731-a612-98009562639c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.360730 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b9h4\" (UniqueName: \"kubernetes.io/projected/86dc24e9-831f-4731-a612-98009562639c-kube-api-access-5b9h4\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq\" (UID: \"86dc24e9-831f-4731-a612-98009562639c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.463232 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86dc24e9-831f-4731-a612-98009562639c-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq\" (UID: \"86dc24e9-831f-4731-a612-98009562639c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.463677 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86dc24e9-831f-4731-a612-98009562639c-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq\" (UID: \"86dc24e9-831f-4731-a612-98009562639c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.463720 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b9h4\" (UniqueName: \"kubernetes.io/projected/86dc24e9-831f-4731-a612-98009562639c-kube-api-access-5b9h4\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq\" (UID: \"86dc24e9-831f-4731-a612-98009562639c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.472772 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86dc24e9-831f-4731-a612-98009562639c-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq\" (UID: \"86dc24e9-831f-4731-a612-98009562639c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.476332 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86dc24e9-831f-4731-a612-98009562639c-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq\" (UID: \"86dc24e9-831f-4731-a612-98009562639c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.482293 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b9h4\" (UniqueName: \"kubernetes.io/projected/86dc24e9-831f-4731-a612-98009562639c-kube-api-access-5b9h4\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq\" (UID: \"86dc24e9-831f-4731-a612-98009562639c\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" Dec 03 06:54:50 crc kubenswrapper[4818]: I1203 06:54:50.682549 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" Dec 03 06:54:51 crc kubenswrapper[4818]: I1203 06:54:51.237516 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ndx9" event={"ID":"6276f22e-c3ff-4d6c-a1b2-02119f513101","Type":"ContainerStarted","Data":"84309fb122f1453446fa2ef29f9be8a83f1d11c36f2ad84530568071a49d9215"} Dec 03 06:54:51 crc kubenswrapper[4818]: I1203 06:54:51.280273 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq"] Dec 03 06:54:51 crc kubenswrapper[4818]: I1203 06:54:51.282648 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4ndx9" podStartSLOduration=2.6774298119999997 podStartE2EDuration="5.282614391s" podCreationTimestamp="2025-12-03 06:54:46 +0000 UTC" firstStartedPulling="2025-12-03 06:54:48.195133466 +0000 UTC m=+1645.886742258" lastFinishedPulling="2025-12-03 06:54:50.800318085 +0000 UTC m=+1648.491926837" observedRunningTime="2025-12-03 06:54:51.274016009 +0000 UTC m=+1648.965624771" watchObservedRunningTime="2025-12-03 06:54:51.282614391 +0000 UTC m=+1648.974223143" Dec 03 06:54:52 crc kubenswrapper[4818]: I1203 06:54:52.249086 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" event={"ID":"86dc24e9-831f-4731-a612-98009562639c","Type":"ContainerStarted","Data":"e2d6dfeb3c9329e67113092be147385082b332378a1f206eb04476a7d82ab0de"} Dec 03 06:54:52 crc kubenswrapper[4818]: I1203 06:54:52.249461 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" event={"ID":"86dc24e9-831f-4731-a612-98009562639c","Type":"ContainerStarted","Data":"b99287bcfcce95874e23e295b967defcd337f91f39180804a91f329fb716ce31"} Dec 03 06:54:52 crc kubenswrapper[4818]: I1203 06:54:52.279332 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" podStartSLOduration=1.712821061 podStartE2EDuration="2.279313678s" podCreationTimestamp="2025-12-03 06:54:50 +0000 UTC" firstStartedPulling="2025-12-03 06:54:51.263230402 +0000 UTC m=+1648.954839154" lastFinishedPulling="2025-12-03 06:54:51.829723009 +0000 UTC m=+1649.521331771" observedRunningTime="2025-12-03 06:54:52.268861559 +0000 UTC m=+1649.960470331" watchObservedRunningTime="2025-12-03 06:54:52.279313678 +0000 UTC m=+1649.970922430" Dec 03 06:54:53 crc kubenswrapper[4818]: I1203 06:54:53.842728 4818 scope.go:117] "RemoveContainer" containerID="1dac9236c1acb12ed2e9b466a699c14085a4a6115d3b106889e255449cd77aaf" Dec 03 06:54:53 crc kubenswrapper[4818]: I1203 06:54:53.870899 4818 scope.go:117] "RemoveContainer" containerID="95ecafd87c822e414bd1fc06f5f75f9b48564c46cc8acbcc900d227f6c28519e" Dec 03 06:54:53 crc kubenswrapper[4818]: I1203 06:54:53.887687 4818 scope.go:117] "RemoveContainer" containerID="231700dfd026f94d1588e43d93e8a8a7dce63750aae9af62ec8862c58010836b" Dec 03 06:54:56 crc kubenswrapper[4818]: I1203 06:54:56.490279 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4ndx9" Dec 03 06:54:56 crc kubenswrapper[4818]: I1203 06:54:56.490660 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4ndx9" Dec 03 06:54:56 crc kubenswrapper[4818]: I1203 06:54:56.533699 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4ndx9" Dec 03 06:54:57 crc kubenswrapper[4818]: I1203 06:54:57.378157 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4ndx9" Dec 03 06:54:57 crc kubenswrapper[4818]: I1203 06:54:57.427685 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4ndx9"] Dec 03 06:54:57 crc kubenswrapper[4818]: I1203 06:54:57.738498 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:54:57 crc kubenswrapper[4818]: E1203 06:54:57.738801 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:54:59 crc kubenswrapper[4818]: I1203 06:54:59.324286 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4ndx9" podUID="6276f22e-c3ff-4d6c-a1b2-02119f513101" containerName="registry-server" containerID="cri-o://84309fb122f1453446fa2ef29f9be8a83f1d11c36f2ad84530568071a49d9215" gracePeriod=2 Dec 03 06:54:59 crc kubenswrapper[4818]: I1203 06:54:59.852288 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4ndx9" Dec 03 06:54:59 crc kubenswrapper[4818]: I1203 06:54:59.986273 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5hpc\" (UniqueName: \"kubernetes.io/projected/6276f22e-c3ff-4d6c-a1b2-02119f513101-kube-api-access-k5hpc\") pod \"6276f22e-c3ff-4d6c-a1b2-02119f513101\" (UID: \"6276f22e-c3ff-4d6c-a1b2-02119f513101\") " Dec 03 06:54:59 crc kubenswrapper[4818]: I1203 06:54:59.986326 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6276f22e-c3ff-4d6c-a1b2-02119f513101-catalog-content\") pod \"6276f22e-c3ff-4d6c-a1b2-02119f513101\" (UID: \"6276f22e-c3ff-4d6c-a1b2-02119f513101\") " Dec 03 06:54:59 crc kubenswrapper[4818]: I1203 06:54:59.986493 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6276f22e-c3ff-4d6c-a1b2-02119f513101-utilities\") pod \"6276f22e-c3ff-4d6c-a1b2-02119f513101\" (UID: \"6276f22e-c3ff-4d6c-a1b2-02119f513101\") " Dec 03 06:54:59 crc kubenswrapper[4818]: I1203 06:54:59.988057 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6276f22e-c3ff-4d6c-a1b2-02119f513101-utilities" (OuterVolumeSpecName: "utilities") pod "6276f22e-c3ff-4d6c-a1b2-02119f513101" (UID: "6276f22e-c3ff-4d6c-a1b2-02119f513101"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:54:59 crc kubenswrapper[4818]: I1203 06:54:59.997892 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6276f22e-c3ff-4d6c-a1b2-02119f513101-kube-api-access-k5hpc" (OuterVolumeSpecName: "kube-api-access-k5hpc") pod "6276f22e-c3ff-4d6c-a1b2-02119f513101" (UID: "6276f22e-c3ff-4d6c-a1b2-02119f513101"). InnerVolumeSpecName "kube-api-access-k5hpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.037332 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6276f22e-c3ff-4d6c-a1b2-02119f513101-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6276f22e-c3ff-4d6c-a1b2-02119f513101" (UID: "6276f22e-c3ff-4d6c-a1b2-02119f513101"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.088932 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6276f22e-c3ff-4d6c-a1b2-02119f513101-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.088983 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5hpc\" (UniqueName: \"kubernetes.io/projected/6276f22e-c3ff-4d6c-a1b2-02119f513101-kube-api-access-k5hpc\") on node \"crc\" DevicePath \"\"" Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.088999 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6276f22e-c3ff-4d6c-a1b2-02119f513101-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.335743 4818 generic.go:334] "Generic (PLEG): container finished" podID="6276f22e-c3ff-4d6c-a1b2-02119f513101" containerID="84309fb122f1453446fa2ef29f9be8a83f1d11c36f2ad84530568071a49d9215" exitCode=0 Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.335785 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ndx9" event={"ID":"6276f22e-c3ff-4d6c-a1b2-02119f513101","Type":"ContainerDied","Data":"84309fb122f1453446fa2ef29f9be8a83f1d11c36f2ad84530568071a49d9215"} Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.335828 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ndx9" event={"ID":"6276f22e-c3ff-4d6c-a1b2-02119f513101","Type":"ContainerDied","Data":"65817d01efafdcc1ee47ea9d9f9358f503d41725b960dae4d34f95a28b2de8f4"} Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.335848 4818 scope.go:117] "RemoveContainer" containerID="84309fb122f1453446fa2ef29f9be8a83f1d11c36f2ad84530568071a49d9215" Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.335872 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4ndx9" Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.373375 4818 scope.go:117] "RemoveContainer" containerID="2a1c3436935aa1aad8fcd2df436373a98c3a928a0ab3ac6dc765312838671b51" Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.442463 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4ndx9"] Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.457274 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4ndx9"] Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.469805 4818 scope.go:117] "RemoveContainer" containerID="4bc7487abd396c9bb7c25deb1830a45c6d7357828f002f2b410da52f563b75a6" Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.510928 4818 scope.go:117] "RemoveContainer" containerID="84309fb122f1453446fa2ef29f9be8a83f1d11c36f2ad84530568071a49d9215" Dec 03 06:55:00 crc kubenswrapper[4818]: E1203 06:55:00.512097 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84309fb122f1453446fa2ef29f9be8a83f1d11c36f2ad84530568071a49d9215\": container with ID starting with 84309fb122f1453446fa2ef29f9be8a83f1d11c36f2ad84530568071a49d9215 not found: ID does not exist" containerID="84309fb122f1453446fa2ef29f9be8a83f1d11c36f2ad84530568071a49d9215" Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.512179 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84309fb122f1453446fa2ef29f9be8a83f1d11c36f2ad84530568071a49d9215"} err="failed to get container status \"84309fb122f1453446fa2ef29f9be8a83f1d11c36f2ad84530568071a49d9215\": rpc error: code = NotFound desc = could not find container \"84309fb122f1453446fa2ef29f9be8a83f1d11c36f2ad84530568071a49d9215\": container with ID starting with 84309fb122f1453446fa2ef29f9be8a83f1d11c36f2ad84530568071a49d9215 not found: ID does not exist" Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.512212 4818 scope.go:117] "RemoveContainer" containerID="2a1c3436935aa1aad8fcd2df436373a98c3a928a0ab3ac6dc765312838671b51" Dec 03 06:55:00 crc kubenswrapper[4818]: E1203 06:55:00.512684 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a1c3436935aa1aad8fcd2df436373a98c3a928a0ab3ac6dc765312838671b51\": container with ID starting with 2a1c3436935aa1aad8fcd2df436373a98c3a928a0ab3ac6dc765312838671b51 not found: ID does not exist" containerID="2a1c3436935aa1aad8fcd2df436373a98c3a928a0ab3ac6dc765312838671b51" Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.512711 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a1c3436935aa1aad8fcd2df436373a98c3a928a0ab3ac6dc765312838671b51"} err="failed to get container status \"2a1c3436935aa1aad8fcd2df436373a98c3a928a0ab3ac6dc765312838671b51\": rpc error: code = NotFound desc = could not find container \"2a1c3436935aa1aad8fcd2df436373a98c3a928a0ab3ac6dc765312838671b51\": container with ID starting with 2a1c3436935aa1aad8fcd2df436373a98c3a928a0ab3ac6dc765312838671b51 not found: ID does not exist" Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.512732 4818 scope.go:117] "RemoveContainer" containerID="4bc7487abd396c9bb7c25deb1830a45c6d7357828f002f2b410da52f563b75a6" Dec 03 06:55:00 crc kubenswrapper[4818]: E1203 06:55:00.513110 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bc7487abd396c9bb7c25deb1830a45c6d7357828f002f2b410da52f563b75a6\": container with ID starting with 4bc7487abd396c9bb7c25deb1830a45c6d7357828f002f2b410da52f563b75a6 not found: ID does not exist" containerID="4bc7487abd396c9bb7c25deb1830a45c6d7357828f002f2b410da52f563b75a6" Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.513158 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bc7487abd396c9bb7c25deb1830a45c6d7357828f002f2b410da52f563b75a6"} err="failed to get container status \"4bc7487abd396c9bb7c25deb1830a45c6d7357828f002f2b410da52f563b75a6\": rpc error: code = NotFound desc = could not find container \"4bc7487abd396c9bb7c25deb1830a45c6d7357828f002f2b410da52f563b75a6\": container with ID starting with 4bc7487abd396c9bb7c25deb1830a45c6d7357828f002f2b410da52f563b75a6 not found: ID does not exist" Dec 03 06:55:00 crc kubenswrapper[4818]: I1203 06:55:00.752343 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6276f22e-c3ff-4d6c-a1b2-02119f513101" path="/var/lib/kubelet/pods/6276f22e-c3ff-4d6c-a1b2-02119f513101/volumes" Dec 03 06:55:04 crc kubenswrapper[4818]: I1203 06:55:04.058344 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-x7rhg"] Dec 03 06:55:04 crc kubenswrapper[4818]: I1203 06:55:04.082512 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-md6rj"] Dec 03 06:55:04 crc kubenswrapper[4818]: I1203 06:55:04.105877 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-15a6-account-create-update-rvmpz"] Dec 03 06:55:04 crc kubenswrapper[4818]: I1203 06:55:04.114350 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-x7rhg"] Dec 03 06:55:04 crc kubenswrapper[4818]: I1203 06:55:04.123588 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-md6rj"] Dec 03 06:55:04 crc kubenswrapper[4818]: I1203 06:55:04.133165 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-15a6-account-create-update-rvmpz"] Dec 03 06:55:04 crc kubenswrapper[4818]: I1203 06:55:04.748844 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19cc3001-7257-4267-a3e2-822ebf77b969" path="/var/lib/kubelet/pods/19cc3001-7257-4267-a3e2-822ebf77b969/volumes" Dec 03 06:55:04 crc kubenswrapper[4818]: I1203 06:55:04.750222 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="619d6969-7ea3-4680-ba99-061f79fcadcc" path="/var/lib/kubelet/pods/619d6969-7ea3-4680-ba99-061f79fcadcc/volumes" Dec 03 06:55:04 crc kubenswrapper[4818]: I1203 06:55:04.750839 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bb6d069-8823-4c5c-be08-d9bbb831e977" path="/var/lib/kubelet/pods/9bb6d069-8823-4c5c-be08-d9bbb831e977/volumes" Dec 03 06:55:06 crc kubenswrapper[4818]: I1203 06:55:06.028005 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-jfgjx"] Dec 03 06:55:06 crc kubenswrapper[4818]: I1203 06:55:06.035443 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-jfgjx"] Dec 03 06:55:06 crc kubenswrapper[4818]: I1203 06:55:06.758529 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68aa86ae-c846-4321-aac9-b37e8837356f" path="/var/lib/kubelet/pods/68aa86ae-c846-4321-aac9-b37e8837356f/volumes" Dec 03 06:55:07 crc kubenswrapper[4818]: I1203 06:55:07.044370 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9db7-account-create-update-6n5md"] Dec 03 06:55:07 crc kubenswrapper[4818]: I1203 06:55:07.056158 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-b465-account-create-update-2bd7d"] Dec 03 06:55:07 crc kubenswrapper[4818]: I1203 06:55:07.067031 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-9db7-account-create-update-6n5md"] Dec 03 06:55:07 crc kubenswrapper[4818]: I1203 06:55:07.078291 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-b465-account-create-update-2bd7d"] Dec 03 06:55:08 crc kubenswrapper[4818]: I1203 06:55:08.751139 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="391197f9-7719-4fae-8698-ea3fe74bc4c0" path="/var/lib/kubelet/pods/391197f9-7719-4fae-8698-ea3fe74bc4c0/volumes" Dec 03 06:55:08 crc kubenswrapper[4818]: I1203 06:55:08.752923 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a7f2204-d32f-4497-b766-b8cbb659bb89" path="/var/lib/kubelet/pods/5a7f2204-d32f-4497-b766-b8cbb659bb89/volumes" Dec 03 06:55:11 crc kubenswrapper[4818]: I1203 06:55:11.738156 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:55:11 crc kubenswrapper[4818]: E1203 06:55:11.738781 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:55:26 crc kubenswrapper[4818]: I1203 06:55:26.737860 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:55:26 crc kubenswrapper[4818]: E1203 06:55:26.738763 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:55:37 crc kubenswrapper[4818]: I1203 06:55:37.738488 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:55:37 crc kubenswrapper[4818]: E1203 06:55:37.739360 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:55:48 crc kubenswrapper[4818]: I1203 06:55:48.062623 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-p57tz"] Dec 03 06:55:48 crc kubenswrapper[4818]: I1203 06:55:48.074838 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-p57tz"] Dec 03 06:55:48 crc kubenswrapper[4818]: I1203 06:55:48.748913 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba5ae97f-71e8-4dc1-bcd4-673e558c9693" path="/var/lib/kubelet/pods/ba5ae97f-71e8-4dc1-bcd4-673e558c9693/volumes" Dec 03 06:55:49 crc kubenswrapper[4818]: I1203 06:55:49.738616 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:55:49 crc kubenswrapper[4818]: E1203 06:55:49.739098 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:55:50 crc kubenswrapper[4818]: I1203 06:55:50.025902 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-wpt88"] Dec 03 06:55:50 crc kubenswrapper[4818]: I1203 06:55:50.033686 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-wpt88"] Dec 03 06:55:50 crc kubenswrapper[4818]: I1203 06:55:50.042461 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-d39d-account-create-update-hjjxx"] Dec 03 06:55:50 crc kubenswrapper[4818]: I1203 06:55:50.050036 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-d39d-account-create-update-hjjxx"] Dec 03 06:55:50 crc kubenswrapper[4818]: I1203 06:55:50.747705 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="388347b6-ce3b-4469-ad87-32faa60f8d21" path="/var/lib/kubelet/pods/388347b6-ce3b-4469-ad87-32faa60f8d21/volumes" Dec 03 06:55:50 crc kubenswrapper[4818]: I1203 06:55:50.748477 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0c0052f-9ec7-43ad-87a0-87a32f8461b7" path="/var/lib/kubelet/pods/e0c0052f-9ec7-43ad-87a0-87a32f8461b7/volumes" Dec 03 06:55:51 crc kubenswrapper[4818]: I1203 06:55:51.029865 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-d7v9h"] Dec 03 06:55:51 crc kubenswrapper[4818]: I1203 06:55:51.039140 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-pdz8d"] Dec 03 06:55:51 crc kubenswrapper[4818]: I1203 06:55:51.048632 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-0a8a-account-create-update-82q4f"] Dec 03 06:55:51 crc kubenswrapper[4818]: I1203 06:55:51.056533 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-d7v9h"] Dec 03 06:55:51 crc kubenswrapper[4818]: I1203 06:55:51.064652 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-0a8a-account-create-update-82q4f"] Dec 03 06:55:51 crc kubenswrapper[4818]: I1203 06:55:51.073071 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-pdz8d"] Dec 03 06:55:52 crc kubenswrapper[4818]: I1203 06:55:52.034435 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-339d-account-create-update-5hpch"] Dec 03 06:55:52 crc kubenswrapper[4818]: I1203 06:55:52.050204 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-339d-account-create-update-5hpch"] Dec 03 06:55:52 crc kubenswrapper[4818]: I1203 06:55:52.751455 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="038a0b6a-32a3-407f-9e31-0f133592cc57" path="/var/lib/kubelet/pods/038a0b6a-32a3-407f-9e31-0f133592cc57/volumes" Dec 03 06:55:52 crc kubenswrapper[4818]: I1203 06:55:52.752881 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d32631e-bf3a-46d9-bfd3-2ea9e21987c0" path="/var/lib/kubelet/pods/4d32631e-bf3a-46d9-bfd3-2ea9e21987c0/volumes" Dec 03 06:55:52 crc kubenswrapper[4818]: I1203 06:55:52.753937 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="939f2f3e-bc03-445c-bc1e-2cb1ca895795" path="/var/lib/kubelet/pods/939f2f3e-bc03-445c-bc1e-2cb1ca895795/volumes" Dec 03 06:55:52 crc kubenswrapper[4818]: I1203 06:55:52.754964 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8063467-b639-4aae-8d0d-db6660dfadac" path="/var/lib/kubelet/pods/c8063467-b639-4aae-8d0d-db6660dfadac/volumes" Dec 03 06:55:53 crc kubenswrapper[4818]: I1203 06:55:53.982066 4818 scope.go:117] "RemoveContainer" containerID="0c0f22c08cd47528c691be66785e371daca73f95c3122d51f2830156c4a7c5a1" Dec 03 06:55:54 crc kubenswrapper[4818]: I1203 06:55:54.024436 4818 scope.go:117] "RemoveContainer" containerID="29688ec029f0eab34d50b274713df996431a8784d7b5298c51eeb3d967467d2c" Dec 03 06:55:54 crc kubenswrapper[4818]: I1203 06:55:54.073149 4818 scope.go:117] "RemoveContainer" containerID="8f6fae7c4ae109b8c787819a2295aefc2e9b087d3e810b8bbcae9c2e838d0762" Dec 03 06:55:54 crc kubenswrapper[4818]: I1203 06:55:54.117179 4818 scope.go:117] "RemoveContainer" containerID="ce9f7385328c736f4ccfa03ca017e311f37ba4f7ac1f4d4eb3c2304ca9011461" Dec 03 06:55:54 crc kubenswrapper[4818]: I1203 06:55:54.180756 4818 scope.go:117] "RemoveContainer" containerID="2a4cea9459d21bb4c38e442530d2daf52454d4c0750a2ae66afe12c7aa793625" Dec 03 06:55:54 crc kubenswrapper[4818]: I1203 06:55:54.216730 4818 scope.go:117] "RemoveContainer" containerID="912cef07c09a6dfc028f919c7b5c63422ad0d9bc0ef5e57df42e8a54d4d56729" Dec 03 06:55:54 crc kubenswrapper[4818]: I1203 06:55:54.325785 4818 scope.go:117] "RemoveContainer" containerID="e915d4dd338656293c4c2bfdb99ff185eee65a4f09098f99f667aac19087fa8f" Dec 03 06:55:54 crc kubenswrapper[4818]: I1203 06:55:54.352557 4818 scope.go:117] "RemoveContainer" containerID="cffcb4d6a6c778c5a1b0f933275c3fa74253fca3bf390eb0e427156485d8b910" Dec 03 06:55:54 crc kubenswrapper[4818]: I1203 06:55:54.395142 4818 scope.go:117] "RemoveContainer" containerID="3a79ad1243ecb7ef8281897632b9168bdae57e2a032e2c042356467c6f110fd2" Dec 03 06:55:54 crc kubenswrapper[4818]: I1203 06:55:54.427825 4818 scope.go:117] "RemoveContainer" containerID="27ead5511efe3d07eb90db2b555cacdba88c01daa96a0a979c13266a88ec897d" Dec 03 06:55:54 crc kubenswrapper[4818]: I1203 06:55:54.459309 4818 scope.go:117] "RemoveContainer" containerID="030d8ac7ea51b4d702857b4ceb17fad2a36177d18d16e2b97820e6a3c67db8c0" Dec 03 06:55:54 crc kubenswrapper[4818]: I1203 06:55:54.483332 4818 scope.go:117] "RemoveContainer" containerID="472f472f6356bdf8aa74bf409dbdc7121ad48d0c2d215e3463d88a85a4f8300f" Dec 03 06:55:54 crc kubenswrapper[4818]: I1203 06:55:54.508093 4818 scope.go:117] "RemoveContainer" containerID="46d79510d850b2af9a38c2776abbeee09cb043e9d83bc8b27b5b905ddcab7942" Dec 03 06:56:02 crc kubenswrapper[4818]: I1203 06:56:02.745869 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:56:02 crc kubenswrapper[4818]: E1203 06:56:02.746863 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:56:04 crc kubenswrapper[4818]: I1203 06:56:04.037412 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-p6k8k"] Dec 03 06:56:04 crc kubenswrapper[4818]: I1203 06:56:04.046905 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-p6k8k"] Dec 03 06:56:04 crc kubenswrapper[4818]: I1203 06:56:04.758520 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed38224f-53f7-4e63-8841-4d1c3f65c456" path="/var/lib/kubelet/pods/ed38224f-53f7-4e63-8841-4d1c3f65c456/volumes" Dec 03 06:56:14 crc kubenswrapper[4818]: I1203 06:56:14.737570 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:56:14 crc kubenswrapper[4818]: E1203 06:56:14.738380 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:56:25 crc kubenswrapper[4818]: I1203 06:56:25.737745 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:56:25 crc kubenswrapper[4818]: E1203 06:56:25.738559 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:56:36 crc kubenswrapper[4818]: I1203 06:56:36.046481 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-ffzld"] Dec 03 06:56:36 crc kubenswrapper[4818]: I1203 06:56:36.058229 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-ffzld"] Dec 03 06:56:36 crc kubenswrapper[4818]: I1203 06:56:36.752303 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7" path="/var/lib/kubelet/pods/cd5f93c7-d706-47d7-8446-7e1b3e1f7ae7/volumes" Dec 03 06:56:38 crc kubenswrapper[4818]: I1203 06:56:38.361461 4818 generic.go:334] "Generic (PLEG): container finished" podID="86dc24e9-831f-4731-a612-98009562639c" containerID="e2d6dfeb3c9329e67113092be147385082b332378a1f206eb04476a7d82ab0de" exitCode=0 Dec 03 06:56:38 crc kubenswrapper[4818]: I1203 06:56:38.361569 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" event={"ID":"86dc24e9-831f-4731-a612-98009562639c","Type":"ContainerDied","Data":"e2d6dfeb3c9329e67113092be147385082b332378a1f206eb04476a7d82ab0de"} Dec 03 06:56:38 crc kubenswrapper[4818]: I1203 06:56:38.738450 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:56:38 crc kubenswrapper[4818]: E1203 06:56:38.738742 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:56:39 crc kubenswrapper[4818]: I1203 06:56:39.888920 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.034799 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-lqzrp"] Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.046461 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-lqzrp"] Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.082842 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86dc24e9-831f-4731-a612-98009562639c-ssh-key\") pod \"86dc24e9-831f-4731-a612-98009562639c\" (UID: \"86dc24e9-831f-4731-a612-98009562639c\") " Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.082953 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86dc24e9-831f-4731-a612-98009562639c-inventory\") pod \"86dc24e9-831f-4731-a612-98009562639c\" (UID: \"86dc24e9-831f-4731-a612-98009562639c\") " Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.083030 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5b9h4\" (UniqueName: \"kubernetes.io/projected/86dc24e9-831f-4731-a612-98009562639c-kube-api-access-5b9h4\") pod \"86dc24e9-831f-4731-a612-98009562639c\" (UID: \"86dc24e9-831f-4731-a612-98009562639c\") " Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.088626 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86dc24e9-831f-4731-a612-98009562639c-kube-api-access-5b9h4" (OuterVolumeSpecName: "kube-api-access-5b9h4") pod "86dc24e9-831f-4731-a612-98009562639c" (UID: "86dc24e9-831f-4731-a612-98009562639c"). InnerVolumeSpecName "kube-api-access-5b9h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.114979 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86dc24e9-831f-4731-a612-98009562639c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "86dc24e9-831f-4731-a612-98009562639c" (UID: "86dc24e9-831f-4731-a612-98009562639c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.117603 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86dc24e9-831f-4731-a612-98009562639c-inventory" (OuterVolumeSpecName: "inventory") pod "86dc24e9-831f-4731-a612-98009562639c" (UID: "86dc24e9-831f-4731-a612-98009562639c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.185379 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/86dc24e9-831f-4731-a612-98009562639c-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.185429 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5b9h4\" (UniqueName: \"kubernetes.io/projected/86dc24e9-831f-4731-a612-98009562639c-kube-api-access-5b9h4\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.185442 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/86dc24e9-831f-4731-a612-98009562639c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.387582 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" event={"ID":"86dc24e9-831f-4731-a612-98009562639c","Type":"ContainerDied","Data":"b99287bcfcce95874e23e295b967defcd337f91f39180804a91f329fb716ce31"} Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.387633 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b99287bcfcce95874e23e295b967defcd337f91f39180804a91f329fb716ce31" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.387673 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.497646 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn"] Dec 03 06:56:40 crc kubenswrapper[4818]: E1203 06:56:40.498228 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6276f22e-c3ff-4d6c-a1b2-02119f513101" containerName="extract-content" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.498257 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="6276f22e-c3ff-4d6c-a1b2-02119f513101" containerName="extract-content" Dec 03 06:56:40 crc kubenswrapper[4818]: E1203 06:56:40.498272 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86dc24e9-831f-4731-a612-98009562639c" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.498284 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="86dc24e9-831f-4731-a612-98009562639c" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 06:56:40 crc kubenswrapper[4818]: E1203 06:56:40.498305 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6276f22e-c3ff-4d6c-a1b2-02119f513101" containerName="extract-utilities" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.498314 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="6276f22e-c3ff-4d6c-a1b2-02119f513101" containerName="extract-utilities" Dec 03 06:56:40 crc kubenswrapper[4818]: E1203 06:56:40.498340 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6276f22e-c3ff-4d6c-a1b2-02119f513101" containerName="registry-server" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.498350 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="6276f22e-c3ff-4d6c-a1b2-02119f513101" containerName="registry-server" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.498601 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="86dc24e9-831f-4731-a612-98009562639c" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.498655 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="6276f22e-c3ff-4d6c-a1b2-02119f513101" containerName="registry-server" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.499480 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.504423 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.504584 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.504675 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2bpsv" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.504761 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.518440 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn"] Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.593531 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71c50082-f6aa-4099-be63-2b33234f0636-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn\" (UID: \"71c50082-f6aa-4099-be63-2b33234f0636\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.593594 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbfhj\" (UniqueName: \"kubernetes.io/projected/71c50082-f6aa-4099-be63-2b33234f0636-kube-api-access-lbfhj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn\" (UID: \"71c50082-f6aa-4099-be63-2b33234f0636\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.593664 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/71c50082-f6aa-4099-be63-2b33234f0636-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn\" (UID: \"71c50082-f6aa-4099-be63-2b33234f0636\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.694932 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71c50082-f6aa-4099-be63-2b33234f0636-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn\" (UID: \"71c50082-f6aa-4099-be63-2b33234f0636\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.695162 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbfhj\" (UniqueName: \"kubernetes.io/projected/71c50082-f6aa-4099-be63-2b33234f0636-kube-api-access-lbfhj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn\" (UID: \"71c50082-f6aa-4099-be63-2b33234f0636\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.695277 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/71c50082-f6aa-4099-be63-2b33234f0636-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn\" (UID: \"71c50082-f6aa-4099-be63-2b33234f0636\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.705470 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/71c50082-f6aa-4099-be63-2b33234f0636-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn\" (UID: \"71c50082-f6aa-4099-be63-2b33234f0636\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.710464 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71c50082-f6aa-4099-be63-2b33234f0636-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn\" (UID: \"71c50082-f6aa-4099-be63-2b33234f0636\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.712673 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbfhj\" (UniqueName: \"kubernetes.io/projected/71c50082-f6aa-4099-be63-2b33234f0636-kube-api-access-lbfhj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn\" (UID: \"71c50082-f6aa-4099-be63-2b33234f0636\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.750964 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c23baebd-4d73-4972-b9eb-5c573e859977" path="/var/lib/kubelet/pods/c23baebd-4d73-4972-b9eb-5c573e859977/volumes" Dec 03 06:56:40 crc kubenswrapper[4818]: I1203 06:56:40.822047 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" Dec 03 06:56:41 crc kubenswrapper[4818]: I1203 06:56:41.363439 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn"] Dec 03 06:56:41 crc kubenswrapper[4818]: I1203 06:56:41.397841 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" event={"ID":"71c50082-f6aa-4099-be63-2b33234f0636","Type":"ContainerStarted","Data":"bf4205a27898aa5b2db76822db218fc724cb6d48ad8f7988d120770a8a53ce11"} Dec 03 06:56:42 crc kubenswrapper[4818]: I1203 06:56:42.413769 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" event={"ID":"71c50082-f6aa-4099-be63-2b33234f0636","Type":"ContainerStarted","Data":"4d37fcb10fd392d87aed80b3f81e816167c07b3a6bae757d238bee4ac6fa8f98"} Dec 03 06:56:42 crc kubenswrapper[4818]: I1203 06:56:42.457325 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" podStartSLOduration=1.889510385 podStartE2EDuration="2.457289334s" podCreationTimestamp="2025-12-03 06:56:40 +0000 UTC" firstStartedPulling="2025-12-03 06:56:41.377839767 +0000 UTC m=+1759.069448529" lastFinishedPulling="2025-12-03 06:56:41.945618706 +0000 UTC m=+1759.637227478" observedRunningTime="2025-12-03 06:56:42.439147925 +0000 UTC m=+1760.130756677" watchObservedRunningTime="2025-12-03 06:56:42.457289334 +0000 UTC m=+1760.148898086" Dec 03 06:56:50 crc kubenswrapper[4818]: I1203 06:56:50.100389 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-v8jgr"] Dec 03 06:56:50 crc kubenswrapper[4818]: I1203 06:56:50.112965 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-v8jgr"] Dec 03 06:56:50 crc kubenswrapper[4818]: I1203 06:56:50.768260 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56e81897-8229-4c21-9beb-a7b0941edf03" path="/var/lib/kubelet/pods/56e81897-8229-4c21-9beb-a7b0941edf03/volumes" Dec 03 06:56:51 crc kubenswrapper[4818]: I1203 06:56:51.738280 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:56:51 crc kubenswrapper[4818]: E1203 06:56:51.738655 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:56:54 crc kubenswrapper[4818]: I1203 06:56:54.820285 4818 scope.go:117] "RemoveContainer" containerID="19a34de912c8415893becf259c87b4bece98c0984dc27039b91e46ed8aeb25b5" Dec 03 06:56:54 crc kubenswrapper[4818]: I1203 06:56:54.877594 4818 scope.go:117] "RemoveContainer" containerID="8805d0d2e882a8ac0a8386c703ab84d3e971b8b279955254aac512e5a53e5d58" Dec 03 06:56:54 crc kubenswrapper[4818]: I1203 06:56:54.914970 4818 scope.go:117] "RemoveContainer" containerID="a20b9ba01523bf083f62485b116ef7085bb29d10e0791cfce5c65d1bbc2107f9" Dec 03 06:56:54 crc kubenswrapper[4818]: I1203 06:56:54.947467 4818 scope.go:117] "RemoveContainer" containerID="a1d52d0e41b793fc88a51dc145961cc2a6b3b6f46b3ca4f38a0e30404c4e6cdf" Dec 03 06:56:56 crc kubenswrapper[4818]: I1203 06:56:56.032777 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-7kbvz"] Dec 03 06:56:56 crc kubenswrapper[4818]: I1203 06:56:56.043311 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-7kbvz"] Dec 03 06:56:56 crc kubenswrapper[4818]: I1203 06:56:56.755955 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="982110da-ac09-4df8-9729-77284ed7539b" path="/var/lib/kubelet/pods/982110da-ac09-4df8-9729-77284ed7539b/volumes" Dec 03 06:57:03 crc kubenswrapper[4818]: I1203 06:57:03.038037 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-978k5"] Dec 03 06:57:03 crc kubenswrapper[4818]: I1203 06:57:03.049258 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-978k5"] Dec 03 06:57:03 crc kubenswrapper[4818]: I1203 06:57:03.738345 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:57:03 crc kubenswrapper[4818]: E1203 06:57:03.738964 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:57:04 crc kubenswrapper[4818]: I1203 06:57:04.750430 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3" path="/var/lib/kubelet/pods/20f94a21-4fd4-43fd-a8d6-0105e6d1c5c3/volumes" Dec 03 06:57:17 crc kubenswrapper[4818]: I1203 06:57:17.737583 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:57:17 crc kubenswrapper[4818]: E1203 06:57:17.738413 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:57:32 crc kubenswrapper[4818]: I1203 06:57:32.745250 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:57:32 crc kubenswrapper[4818]: E1203 06:57:32.746143 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:57:44 crc kubenswrapper[4818]: I1203 06:57:44.058193 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-sb9lv"] Dec 03 06:57:44 crc kubenswrapper[4818]: I1203 06:57:44.071508 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-vn2ml"] Dec 03 06:57:44 crc kubenswrapper[4818]: I1203 06:57:44.107535 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-gff9c"] Dec 03 06:57:44 crc kubenswrapper[4818]: I1203 06:57:44.118127 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-e5ab-account-create-update-n6pl8"] Dec 03 06:57:44 crc kubenswrapper[4818]: I1203 06:57:44.142045 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-5631-account-create-update-k87bh"] Dec 03 06:57:44 crc kubenswrapper[4818]: I1203 06:57:44.149658 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-gff9c"] Dec 03 06:57:44 crc kubenswrapper[4818]: I1203 06:57:44.156721 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-5631-account-create-update-k87bh"] Dec 03 06:57:44 crc kubenswrapper[4818]: I1203 06:57:44.163772 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-e5ab-account-create-update-n6pl8"] Dec 03 06:57:44 crc kubenswrapper[4818]: I1203 06:57:44.171286 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-vn2ml"] Dec 03 06:57:44 crc kubenswrapper[4818]: I1203 06:57:44.178321 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-sb9lv"] Dec 03 06:57:44 crc kubenswrapper[4818]: I1203 06:57:44.758252 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dcf3310-6518-4c25-a81c-ef30724db1b8" path="/var/lib/kubelet/pods/4dcf3310-6518-4c25-a81c-ef30724db1b8/volumes" Dec 03 06:57:44 crc kubenswrapper[4818]: I1203 06:57:44.759563 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5184be52-db4d-4e95-88bd-a6569ddb75ff" path="/var/lib/kubelet/pods/5184be52-db4d-4e95-88bd-a6569ddb75ff/volumes" Dec 03 06:57:44 crc kubenswrapper[4818]: I1203 06:57:44.760794 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae575e67-4d14-4f4c-8d27-b74f9b837134" path="/var/lib/kubelet/pods/ae575e67-4d14-4f4c-8d27-b74f9b837134/volumes" Dec 03 06:57:44 crc kubenswrapper[4818]: I1203 06:57:44.762079 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2194aad-43ff-4ab6-b648-e079a0688a5f" path="/var/lib/kubelet/pods/f2194aad-43ff-4ab6-b648-e079a0688a5f/volumes" Dec 03 06:57:44 crc kubenswrapper[4818]: I1203 06:57:44.764325 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7514750-68e9-462c-b401-f80689f8f4df" path="/var/lib/kubelet/pods/f7514750-68e9-462c-b401-f80689f8f4df/volumes" Dec 03 06:57:45 crc kubenswrapper[4818]: I1203 06:57:45.042490 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-8ffc-account-create-update-s65jt"] Dec 03 06:57:45 crc kubenswrapper[4818]: I1203 06:57:45.054681 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-8ffc-account-create-update-s65jt"] Dec 03 06:57:46 crc kubenswrapper[4818]: I1203 06:57:46.749138 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f0e2585-56c1-435c-9db7-74b19a1cab80" path="/var/lib/kubelet/pods/7f0e2585-56c1-435c-9db7-74b19a1cab80/volumes" Dec 03 06:57:47 crc kubenswrapper[4818]: I1203 06:57:47.738020 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:57:47 crc kubenswrapper[4818]: E1203 06:57:47.738483 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:57:55 crc kubenswrapper[4818]: I1203 06:57:55.075377 4818 scope.go:117] "RemoveContainer" containerID="c02300f0d9f614a680c6a19b0d6e6dddc1cd095065722b3b46911227341903fa" Dec 03 06:57:55 crc kubenswrapper[4818]: I1203 06:57:55.110833 4818 scope.go:117] "RemoveContainer" containerID="40a20a4a921d35cbe1c425b32780cb643b7e62e79caade8cedeed1463cf89cf3" Dec 03 06:57:55 crc kubenswrapper[4818]: I1203 06:57:55.178949 4818 scope.go:117] "RemoveContainer" containerID="74684d24944f4d5aed5bbb34bd9cd0d54f3d12dc4525efa4bb85050dd7694d97" Dec 03 06:57:55 crc kubenswrapper[4818]: I1203 06:57:55.229563 4818 scope.go:117] "RemoveContainer" containerID="8ff182159e975cf1a6bd5f92183c3e16693a90e305a4a18dca48b23710bd8cee" Dec 03 06:57:55 crc kubenswrapper[4818]: I1203 06:57:55.274607 4818 scope.go:117] "RemoveContainer" containerID="2eec35e25d66cc807661730864b68b790a563c984227ff43593f4242c9b2ed83" Dec 03 06:57:55 crc kubenswrapper[4818]: I1203 06:57:55.342258 4818 scope.go:117] "RemoveContainer" containerID="c38ac13a89962c7d38a2a09e6bc8c242c29b7d8e01fba581a8988df80d6820fd" Dec 03 06:57:55 crc kubenswrapper[4818]: I1203 06:57:55.362010 4818 scope.go:117] "RemoveContainer" containerID="2051b0fea1f9046bee62c7be1d04a2562c80f88640717a85b180d8513376c4e0" Dec 03 06:57:55 crc kubenswrapper[4818]: I1203 06:57:55.385490 4818 scope.go:117] "RemoveContainer" containerID="7aa75430e5eda89fc452183734b53c46d491e15429f34cbebdaf233a74322f2c" Dec 03 06:58:00 crc kubenswrapper[4818]: I1203 06:58:00.278165 4818 generic.go:334] "Generic (PLEG): container finished" podID="71c50082-f6aa-4099-be63-2b33234f0636" containerID="4d37fcb10fd392d87aed80b3f81e816167c07b3a6bae757d238bee4ac6fa8f98" exitCode=0 Dec 03 06:58:00 crc kubenswrapper[4818]: I1203 06:58:00.278372 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" event={"ID":"71c50082-f6aa-4099-be63-2b33234f0636","Type":"ContainerDied","Data":"4d37fcb10fd392d87aed80b3f81e816167c07b3a6bae757d238bee4ac6fa8f98"} Dec 03 06:58:01 crc kubenswrapper[4818]: I1203 06:58:01.717533 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" Dec 03 06:58:01 crc kubenswrapper[4818]: I1203 06:58:01.784774 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71c50082-f6aa-4099-be63-2b33234f0636-inventory\") pod \"71c50082-f6aa-4099-be63-2b33234f0636\" (UID: \"71c50082-f6aa-4099-be63-2b33234f0636\") " Dec 03 06:58:01 crc kubenswrapper[4818]: I1203 06:58:01.785083 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/71c50082-f6aa-4099-be63-2b33234f0636-ssh-key\") pod \"71c50082-f6aa-4099-be63-2b33234f0636\" (UID: \"71c50082-f6aa-4099-be63-2b33234f0636\") " Dec 03 06:58:01 crc kubenswrapper[4818]: I1203 06:58:01.785171 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbfhj\" (UniqueName: \"kubernetes.io/projected/71c50082-f6aa-4099-be63-2b33234f0636-kube-api-access-lbfhj\") pod \"71c50082-f6aa-4099-be63-2b33234f0636\" (UID: \"71c50082-f6aa-4099-be63-2b33234f0636\") " Dec 03 06:58:01 crc kubenswrapper[4818]: I1203 06:58:01.798597 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71c50082-f6aa-4099-be63-2b33234f0636-kube-api-access-lbfhj" (OuterVolumeSpecName: "kube-api-access-lbfhj") pod "71c50082-f6aa-4099-be63-2b33234f0636" (UID: "71c50082-f6aa-4099-be63-2b33234f0636"). InnerVolumeSpecName "kube-api-access-lbfhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:58:01 crc kubenswrapper[4818]: I1203 06:58:01.817138 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71c50082-f6aa-4099-be63-2b33234f0636-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "71c50082-f6aa-4099-be63-2b33234f0636" (UID: "71c50082-f6aa-4099-be63-2b33234f0636"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:58:01 crc kubenswrapper[4818]: I1203 06:58:01.817560 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71c50082-f6aa-4099-be63-2b33234f0636-inventory" (OuterVolumeSpecName: "inventory") pod "71c50082-f6aa-4099-be63-2b33234f0636" (UID: "71c50082-f6aa-4099-be63-2b33234f0636"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:58:01 crc kubenswrapper[4818]: I1203 06:58:01.887269 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/71c50082-f6aa-4099-be63-2b33234f0636-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 06:58:01 crc kubenswrapper[4818]: I1203 06:58:01.887301 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbfhj\" (UniqueName: \"kubernetes.io/projected/71c50082-f6aa-4099-be63-2b33234f0636-kube-api-access-lbfhj\") on node \"crc\" DevicePath \"\"" Dec 03 06:58:01 crc kubenswrapper[4818]: I1203 06:58:01.887311 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71c50082-f6aa-4099-be63-2b33234f0636-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.304161 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" event={"ID":"71c50082-f6aa-4099-be63-2b33234f0636","Type":"ContainerDied","Data":"bf4205a27898aa5b2db76822db218fc724cb6d48ad8f7988d120770a8a53ce11"} Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.304461 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf4205a27898aa5b2db76822db218fc724cb6d48ad8f7988d120770a8a53ce11" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.304226 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.456018 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh"] Dec 03 06:58:02 crc kubenswrapper[4818]: E1203 06:58:02.456679 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71c50082-f6aa-4099-be63-2b33234f0636" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.456712 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="71c50082-f6aa-4099-be63-2b33234f0636" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.457108 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="71c50082-f6aa-4099-be63-2b33234f0636" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.458170 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.460471 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.461072 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.462580 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.463468 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2bpsv" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.471572 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh"] Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.501672 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh\" (UID: \"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.501801 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh\" (UID: \"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.501928 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ht4d\" (UniqueName: \"kubernetes.io/projected/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-kube-api-access-2ht4d\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh\" (UID: \"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.603180 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ht4d\" (UniqueName: \"kubernetes.io/projected/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-kube-api-access-2ht4d\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh\" (UID: \"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.603566 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh\" (UID: \"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.603754 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh\" (UID: \"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.611391 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh\" (UID: \"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.611497 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh\" (UID: \"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.621952 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ht4d\" (UniqueName: \"kubernetes.io/projected/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-kube-api-access-2ht4d\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh\" (UID: \"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.749710 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:58:02 crc kubenswrapper[4818]: E1203 06:58:02.749983 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 06:58:02 crc kubenswrapper[4818]: I1203 06:58:02.784951 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" Dec 03 06:58:03 crc kubenswrapper[4818]: I1203 06:58:03.359783 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh"] Dec 03 06:58:04 crc kubenswrapper[4818]: I1203 06:58:04.322794 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" event={"ID":"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2","Type":"ContainerStarted","Data":"32f4d3908ccc63a21548b437cabc8535dcbd5200cad175463e258644bebbc8b1"} Dec 03 06:58:04 crc kubenswrapper[4818]: I1203 06:58:04.323120 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" event={"ID":"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2","Type":"ContainerStarted","Data":"b8b794f755313d4e7fbb9c5e9e0591dc0d66e4b3e63faadf7bf8e59fec9cd30c"} Dec 03 06:58:04 crc kubenswrapper[4818]: I1203 06:58:04.354064 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" podStartSLOduration=1.8601813200000001 podStartE2EDuration="2.354041527s" podCreationTimestamp="2025-12-03 06:58:02 +0000 UTC" firstStartedPulling="2025-12-03 06:58:03.367653706 +0000 UTC m=+1841.059262458" lastFinishedPulling="2025-12-03 06:58:03.861513913 +0000 UTC m=+1841.553122665" observedRunningTime="2025-12-03 06:58:04.344501021 +0000 UTC m=+1842.036109803" watchObservedRunningTime="2025-12-03 06:58:04.354041527 +0000 UTC m=+1842.045650279" Dec 03 06:58:09 crc kubenswrapper[4818]: I1203 06:58:09.369883 4818 generic.go:334] "Generic (PLEG): container finished" podID="4e2bf3bf-bd96-4042-b1e6-11247aefdbe2" containerID="32f4d3908ccc63a21548b437cabc8535dcbd5200cad175463e258644bebbc8b1" exitCode=0 Dec 03 06:58:09 crc kubenswrapper[4818]: I1203 06:58:09.370135 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" event={"ID":"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2","Type":"ContainerDied","Data":"32f4d3908ccc63a21548b437cabc8535dcbd5200cad175463e258644bebbc8b1"} Dec 03 06:58:10 crc kubenswrapper[4818]: I1203 06:58:10.765842 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" Dec 03 06:58:10 crc kubenswrapper[4818]: I1203 06:58:10.904372 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-inventory\") pod \"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2\" (UID: \"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2\") " Dec 03 06:58:10 crc kubenswrapper[4818]: I1203 06:58:10.904685 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ht4d\" (UniqueName: \"kubernetes.io/projected/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-kube-api-access-2ht4d\") pod \"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2\" (UID: \"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2\") " Dec 03 06:58:10 crc kubenswrapper[4818]: I1203 06:58:10.904790 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-ssh-key\") pod \"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2\" (UID: \"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2\") " Dec 03 06:58:10 crc kubenswrapper[4818]: I1203 06:58:10.914308 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-kube-api-access-2ht4d" (OuterVolumeSpecName: "kube-api-access-2ht4d") pod "4e2bf3bf-bd96-4042-b1e6-11247aefdbe2" (UID: "4e2bf3bf-bd96-4042-b1e6-11247aefdbe2"). InnerVolumeSpecName "kube-api-access-2ht4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:58:10 crc kubenswrapper[4818]: I1203 06:58:10.939951 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4e2bf3bf-bd96-4042-b1e6-11247aefdbe2" (UID: "4e2bf3bf-bd96-4042-b1e6-11247aefdbe2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:58:10 crc kubenswrapper[4818]: I1203 06:58:10.963787 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-inventory" (OuterVolumeSpecName: "inventory") pod "4e2bf3bf-bd96-4042-b1e6-11247aefdbe2" (UID: "4e2bf3bf-bd96-4042-b1e6-11247aefdbe2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.006385 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.006424 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.006437 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ht4d\" (UniqueName: \"kubernetes.io/projected/4e2bf3bf-bd96-4042-b1e6-11247aefdbe2-kube-api-access-2ht4d\") on node \"crc\" DevicePath \"\"" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.392736 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" event={"ID":"4e2bf3bf-bd96-4042-b1e6-11247aefdbe2","Type":"ContainerDied","Data":"b8b794f755313d4e7fbb9c5e9e0591dc0d66e4b3e63faadf7bf8e59fec9cd30c"} Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.392793 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8b794f755313d4e7fbb9c5e9e0591dc0d66e4b3e63faadf7bf8e59fec9cd30c" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.392794 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.491201 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4"] Dec 03 06:58:11 crc kubenswrapper[4818]: E1203 06:58:11.491660 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e2bf3bf-bd96-4042-b1e6-11247aefdbe2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.491675 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e2bf3bf-bd96-4042-b1e6-11247aefdbe2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.491894 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e2bf3bf-bd96-4042-b1e6-11247aefdbe2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.492610 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.494931 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.496372 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2bpsv" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.496655 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.496890 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.508746 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4"] Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.617423 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdfz4\" (UID: \"9568e2c6-4f91-49ce-80e6-0081b7bef2c5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.617535 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdfz4\" (UID: \"9568e2c6-4f91-49ce-80e6-0081b7bef2c5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.617721 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl58b\" (UniqueName: \"kubernetes.io/projected/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-kube-api-access-dl58b\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdfz4\" (UID: \"9568e2c6-4f91-49ce-80e6-0081b7bef2c5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.719700 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdfz4\" (UID: \"9568e2c6-4f91-49ce-80e6-0081b7bef2c5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.719959 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdfz4\" (UID: \"9568e2c6-4f91-49ce-80e6-0081b7bef2c5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.720029 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl58b\" (UniqueName: \"kubernetes.io/projected/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-kube-api-access-dl58b\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdfz4\" (UID: \"9568e2c6-4f91-49ce-80e6-0081b7bef2c5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.724621 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdfz4\" (UID: \"9568e2c6-4f91-49ce-80e6-0081b7bef2c5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.733860 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdfz4\" (UID: \"9568e2c6-4f91-49ce-80e6-0081b7bef2c5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.740702 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl58b\" (UniqueName: \"kubernetes.io/projected/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-kube-api-access-dl58b\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pdfz4\" (UID: \"9568e2c6-4f91-49ce-80e6-0081b7bef2c5\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" Dec 03 06:58:11 crc kubenswrapper[4818]: I1203 06:58:11.809307 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" Dec 03 06:58:12 crc kubenswrapper[4818]: I1203 06:58:12.362287 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4"] Dec 03 06:58:12 crc kubenswrapper[4818]: I1203 06:58:12.400529 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" event={"ID":"9568e2c6-4f91-49ce-80e6-0081b7bef2c5","Type":"ContainerStarted","Data":"a48cb2003c46a76a0e2b2476bb79bf1b25aa888fc02a97f98b59cce462ed670a"} Dec 03 06:58:13 crc kubenswrapper[4818]: I1203 06:58:13.441014 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" event={"ID":"9568e2c6-4f91-49ce-80e6-0081b7bef2c5","Type":"ContainerStarted","Data":"a5d5f24b2f5503abf49b6f57d5e32d68f95b0f03f1387e70b69ae711f2b29764"} Dec 03 06:58:13 crc kubenswrapper[4818]: I1203 06:58:13.465698 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" podStartSLOduration=2.034197233 podStartE2EDuration="2.465675645s" podCreationTimestamp="2025-12-03 06:58:11 +0000 UTC" firstStartedPulling="2025-12-03 06:58:12.369457923 +0000 UTC m=+1850.061066675" lastFinishedPulling="2025-12-03 06:58:12.800936335 +0000 UTC m=+1850.492545087" observedRunningTime="2025-12-03 06:58:13.455841761 +0000 UTC m=+1851.147450513" watchObservedRunningTime="2025-12-03 06:58:13.465675645 +0000 UTC m=+1851.157284387" Dec 03 06:58:14 crc kubenswrapper[4818]: I1203 06:58:14.738350 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 06:58:15 crc kubenswrapper[4818]: I1203 06:58:15.462073 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"955318249bafbdd3f6163ed23bb008bdbe9f0304f5cc29a8cc76bbceb8f4da5b"} Dec 03 06:58:22 crc kubenswrapper[4818]: I1203 06:58:22.043590 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kxhw2"] Dec 03 06:58:22 crc kubenswrapper[4818]: I1203 06:58:22.053805 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kxhw2"] Dec 03 06:58:22 crc kubenswrapper[4818]: I1203 06:58:22.750661 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987" path="/var/lib/kubelet/pods/c959ad1c-85fd-4cf0-a3a7-1cbf91cc7987/volumes" Dec 03 06:58:45 crc kubenswrapper[4818]: I1203 06:58:45.050024 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-d8cch"] Dec 03 06:58:45 crc kubenswrapper[4818]: I1203 06:58:45.056840 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-ql2nx"] Dec 03 06:58:45 crc kubenswrapper[4818]: I1203 06:58:45.063457 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-d8cch"] Dec 03 06:58:45 crc kubenswrapper[4818]: I1203 06:58:45.070573 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-ql2nx"] Dec 03 06:58:46 crc kubenswrapper[4818]: I1203 06:58:46.750729 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c" path="/var/lib/kubelet/pods/199309f9-ae9c-41cf-bd3a-dadf0f9c1c6c/volumes" Dec 03 06:58:46 crc kubenswrapper[4818]: I1203 06:58:46.751683 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c1214a4-c9ac-42a3-93d9-ac9ac361c544" path="/var/lib/kubelet/pods/9c1214a4-c9ac-42a3-93d9-ac9ac361c544/volumes" Dec 03 06:58:53 crc kubenswrapper[4818]: I1203 06:58:53.847601 4818 generic.go:334] "Generic (PLEG): container finished" podID="9568e2c6-4f91-49ce-80e6-0081b7bef2c5" containerID="a5d5f24b2f5503abf49b6f57d5e32d68f95b0f03f1387e70b69ae711f2b29764" exitCode=0 Dec 03 06:58:53 crc kubenswrapper[4818]: I1203 06:58:53.847701 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" event={"ID":"9568e2c6-4f91-49ce-80e6-0081b7bef2c5","Type":"ContainerDied","Data":"a5d5f24b2f5503abf49b6f57d5e32d68f95b0f03f1387e70b69ae711f2b29764"} Dec 03 06:58:55 crc kubenswrapper[4818]: I1203 06:58:55.278411 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" Dec 03 06:58:55 crc kubenswrapper[4818]: I1203 06:58:55.459283 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl58b\" (UniqueName: \"kubernetes.io/projected/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-kube-api-access-dl58b\") pod \"9568e2c6-4f91-49ce-80e6-0081b7bef2c5\" (UID: \"9568e2c6-4f91-49ce-80e6-0081b7bef2c5\") " Dec 03 06:58:55 crc kubenswrapper[4818]: I1203 06:58:55.459374 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-inventory\") pod \"9568e2c6-4f91-49ce-80e6-0081b7bef2c5\" (UID: \"9568e2c6-4f91-49ce-80e6-0081b7bef2c5\") " Dec 03 06:58:55 crc kubenswrapper[4818]: I1203 06:58:55.459458 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-ssh-key\") pod \"9568e2c6-4f91-49ce-80e6-0081b7bef2c5\" (UID: \"9568e2c6-4f91-49ce-80e6-0081b7bef2c5\") " Dec 03 06:58:55 crc kubenswrapper[4818]: I1203 06:58:55.472076 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-kube-api-access-dl58b" (OuterVolumeSpecName: "kube-api-access-dl58b") pod "9568e2c6-4f91-49ce-80e6-0081b7bef2c5" (UID: "9568e2c6-4f91-49ce-80e6-0081b7bef2c5"). InnerVolumeSpecName "kube-api-access-dl58b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:58:55 crc kubenswrapper[4818]: I1203 06:58:55.494662 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-inventory" (OuterVolumeSpecName: "inventory") pod "9568e2c6-4f91-49ce-80e6-0081b7bef2c5" (UID: "9568e2c6-4f91-49ce-80e6-0081b7bef2c5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:58:55 crc kubenswrapper[4818]: I1203 06:58:55.497542 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9568e2c6-4f91-49ce-80e6-0081b7bef2c5" (UID: "9568e2c6-4f91-49ce-80e6-0081b7bef2c5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:58:55 crc kubenswrapper[4818]: I1203 06:58:55.529296 4818 scope.go:117] "RemoveContainer" containerID="b774aa954ffb4a76150f227eb0d04274bf155df99dc072805eac3f69b14c3b2f" Dec 03 06:58:55 crc kubenswrapper[4818]: I1203 06:58:55.563640 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 06:58:55 crc kubenswrapper[4818]: I1203 06:58:55.563672 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 06:58:55 crc kubenswrapper[4818]: I1203 06:58:55.563685 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl58b\" (UniqueName: \"kubernetes.io/projected/9568e2c6-4f91-49ce-80e6-0081b7bef2c5-kube-api-access-dl58b\") on node \"crc\" DevicePath \"\"" Dec 03 06:58:55 crc kubenswrapper[4818]: I1203 06:58:55.608961 4818 scope.go:117] "RemoveContainer" containerID="3e0177bc27efb929e78580f92516e4c3ff4cc29a316b0119f792dc40a42fd1ae" Dec 03 06:58:55 crc kubenswrapper[4818]: I1203 06:58:55.653247 4818 scope.go:117] "RemoveContainer" containerID="20091f6cfd086f85c0abaf606821fed5cbf8f11a0ce460a565af3f3e2f6f9f0d" Dec 03 06:58:55 crc kubenswrapper[4818]: I1203 06:58:55.867966 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" event={"ID":"9568e2c6-4f91-49ce-80e6-0081b7bef2c5","Type":"ContainerDied","Data":"a48cb2003c46a76a0e2b2476bb79bf1b25aa888fc02a97f98b59cce462ed670a"} Dec 03 06:58:55 crc kubenswrapper[4818]: I1203 06:58:55.868005 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a48cb2003c46a76a0e2b2476bb79bf1b25aa888fc02a97f98b59cce462ed670a" Dec 03 06:58:55 crc kubenswrapper[4818]: I1203 06:58:55.868036 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pdfz4" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.005635 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9"] Dec 03 06:58:56 crc kubenswrapper[4818]: E1203 06:58:56.006053 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9568e2c6-4f91-49ce-80e6-0081b7bef2c5" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.006070 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9568e2c6-4f91-49ce-80e6-0081b7bef2c5" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.006301 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="9568e2c6-4f91-49ce-80e6-0081b7bef2c5" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.007029 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.012004 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2bpsv" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.012253 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.023670 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9"] Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.024362 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.024461 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.175760 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9\" (UID: \"2ef6c961-25b1-4f6f-a98b-65d79787e9e1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.175883 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxgwx\" (UniqueName: \"kubernetes.io/projected/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-kube-api-access-sxgwx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9\" (UID: \"2ef6c961-25b1-4f6f-a98b-65d79787e9e1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.176043 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9\" (UID: \"2ef6c961-25b1-4f6f-a98b-65d79787e9e1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.277582 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9\" (UID: \"2ef6c961-25b1-4f6f-a98b-65d79787e9e1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.277640 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxgwx\" (UniqueName: \"kubernetes.io/projected/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-kube-api-access-sxgwx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9\" (UID: \"2ef6c961-25b1-4f6f-a98b-65d79787e9e1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.277745 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9\" (UID: \"2ef6c961-25b1-4f6f-a98b-65d79787e9e1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.283017 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9\" (UID: \"2ef6c961-25b1-4f6f-a98b-65d79787e9e1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.283993 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9\" (UID: \"2ef6c961-25b1-4f6f-a98b-65d79787e9e1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.293382 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxgwx\" (UniqueName: \"kubernetes.io/projected/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-kube-api-access-sxgwx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9\" (UID: \"2ef6c961-25b1-4f6f-a98b-65d79787e9e1\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.333163 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.848475 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9"] Dec 03 06:58:56 crc kubenswrapper[4818]: I1203 06:58:56.881040 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" event={"ID":"2ef6c961-25b1-4f6f-a98b-65d79787e9e1","Type":"ContainerStarted","Data":"78a7713dfefaa19c640a14dba81c1853eaec9466fd6f03218ca1d60aaad57a78"} Dec 03 06:58:57 crc kubenswrapper[4818]: I1203 06:58:57.893529 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" event={"ID":"2ef6c961-25b1-4f6f-a98b-65d79787e9e1","Type":"ContainerStarted","Data":"40497a74a089b1cc693a25ef31d9bf989ca992914586219a4b1b5ea16d50e69f"} Dec 03 06:58:57 crc kubenswrapper[4818]: I1203 06:58:57.926681 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" podStartSLOduration=2.44307358 podStartE2EDuration="2.926657823s" podCreationTimestamp="2025-12-03 06:58:55 +0000 UTC" firstStartedPulling="2025-12-03 06:58:56.856975898 +0000 UTC m=+1894.548584640" lastFinishedPulling="2025-12-03 06:58:57.340560111 +0000 UTC m=+1895.032168883" observedRunningTime="2025-12-03 06:58:57.914043721 +0000 UTC m=+1895.605652483" watchObservedRunningTime="2025-12-03 06:58:57.926657823 +0000 UTC m=+1895.618266575" Dec 03 06:59:30 crc kubenswrapper[4818]: I1203 06:59:30.053170 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-flwjh"] Dec 03 06:59:30 crc kubenswrapper[4818]: I1203 06:59:30.061623 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-flwjh"] Dec 03 06:59:30 crc kubenswrapper[4818]: I1203 06:59:30.751363 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e23ee654-89e4-449e-8417-23775d4dcbfb" path="/var/lib/kubelet/pods/e23ee654-89e4-449e-8417-23775d4dcbfb/volumes" Dec 03 06:59:51 crc kubenswrapper[4818]: I1203 06:59:51.432539 4818 generic.go:334] "Generic (PLEG): container finished" podID="2ef6c961-25b1-4f6f-a98b-65d79787e9e1" containerID="40497a74a089b1cc693a25ef31d9bf989ca992914586219a4b1b5ea16d50e69f" exitCode=0 Dec 03 06:59:51 crc kubenswrapper[4818]: I1203 06:59:51.432608 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" event={"ID":"2ef6c961-25b1-4f6f-a98b-65d79787e9e1","Type":"ContainerDied","Data":"40497a74a089b1cc693a25ef31d9bf989ca992914586219a4b1b5ea16d50e69f"} Dec 03 06:59:52 crc kubenswrapper[4818]: I1203 06:59:52.824429 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" Dec 03 06:59:52 crc kubenswrapper[4818]: I1203 06:59:52.955760 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-inventory\") pod \"2ef6c961-25b1-4f6f-a98b-65d79787e9e1\" (UID: \"2ef6c961-25b1-4f6f-a98b-65d79787e9e1\") " Dec 03 06:59:52 crc kubenswrapper[4818]: I1203 06:59:52.956027 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxgwx\" (UniqueName: \"kubernetes.io/projected/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-kube-api-access-sxgwx\") pod \"2ef6c961-25b1-4f6f-a98b-65d79787e9e1\" (UID: \"2ef6c961-25b1-4f6f-a98b-65d79787e9e1\") " Dec 03 06:59:52 crc kubenswrapper[4818]: I1203 06:59:52.956520 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-ssh-key\") pod \"2ef6c961-25b1-4f6f-a98b-65d79787e9e1\" (UID: \"2ef6c961-25b1-4f6f-a98b-65d79787e9e1\") " Dec 03 06:59:52 crc kubenswrapper[4818]: I1203 06:59:52.964352 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-kube-api-access-sxgwx" (OuterVolumeSpecName: "kube-api-access-sxgwx") pod "2ef6c961-25b1-4f6f-a98b-65d79787e9e1" (UID: "2ef6c961-25b1-4f6f-a98b-65d79787e9e1"). InnerVolumeSpecName "kube-api-access-sxgwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:52.995551 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2ef6c961-25b1-4f6f-a98b-65d79787e9e1" (UID: "2ef6c961-25b1-4f6f-a98b-65d79787e9e1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.011587 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-inventory" (OuterVolumeSpecName: "inventory") pod "2ef6c961-25b1-4f6f-a98b-65d79787e9e1" (UID: "2ef6c961-25b1-4f6f-a98b-65d79787e9e1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.060075 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxgwx\" (UniqueName: \"kubernetes.io/projected/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-kube-api-access-sxgwx\") on node \"crc\" DevicePath \"\"" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.060121 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.060143 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ef6c961-25b1-4f6f-a98b-65d79787e9e1-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.453350 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" event={"ID":"2ef6c961-25b1-4f6f-a98b-65d79787e9e1","Type":"ContainerDied","Data":"78a7713dfefaa19c640a14dba81c1853eaec9466fd6f03218ca1d60aaad57a78"} Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.453401 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78a7713dfefaa19c640a14dba81c1853eaec9466fd6f03218ca1d60aaad57a78" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.453469 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.532740 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-lbkfx"] Dec 03 06:59:53 crc kubenswrapper[4818]: E1203 06:59:53.533240 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ef6c961-25b1-4f6f-a98b-65d79787e9e1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.533265 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ef6c961-25b1-4f6f-a98b-65d79787e9e1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.533519 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ef6c961-25b1-4f6f-a98b-65d79787e9e1" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.534272 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.539263 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.539347 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2bpsv" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.539533 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.539581 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.544395 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-lbkfx"] Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.692894 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-lbkfx\" (UID: \"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de\") " pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.692984 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-lbkfx\" (UID: \"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de\") " pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.693093 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jm6kc\" (UniqueName: \"kubernetes.io/projected/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-kube-api-access-jm6kc\") pod \"ssh-known-hosts-edpm-deployment-lbkfx\" (UID: \"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de\") " pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.795328 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-lbkfx\" (UID: \"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de\") " pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.795620 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jm6kc\" (UniqueName: \"kubernetes.io/projected/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-kube-api-access-jm6kc\") pod \"ssh-known-hosts-edpm-deployment-lbkfx\" (UID: \"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de\") " pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.795726 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-lbkfx\" (UID: \"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de\") " pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.799077 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-lbkfx\" (UID: \"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de\") " pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.799857 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-lbkfx\" (UID: \"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de\") " pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.819921 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jm6kc\" (UniqueName: \"kubernetes.io/projected/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-kube-api-access-jm6kc\") pod \"ssh-known-hosts-edpm-deployment-lbkfx\" (UID: \"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de\") " pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" Dec 03 06:59:53 crc kubenswrapper[4818]: I1203 06:59:53.851100 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" Dec 03 06:59:54 crc kubenswrapper[4818]: I1203 06:59:54.583130 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-lbkfx"] Dec 03 06:59:54 crc kubenswrapper[4818]: W1203 06:59:54.592596 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb054c1a4_49a8_46eb_9e6d_aaaab7cdf2de.slice/crio-3bc3d91a09348073a9803998c16cbccd9f9f3c00b2788fbafa2b2a8060be7f4d WatchSource:0}: Error finding container 3bc3d91a09348073a9803998c16cbccd9f9f3c00b2788fbafa2b2a8060be7f4d: Status 404 returned error can't find the container with id 3bc3d91a09348073a9803998c16cbccd9f9f3c00b2788fbafa2b2a8060be7f4d Dec 03 06:59:54 crc kubenswrapper[4818]: I1203 06:59:54.594594 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 06:59:55 crc kubenswrapper[4818]: I1203 06:59:55.559539 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" event={"ID":"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de","Type":"ContainerStarted","Data":"18213a6eecd26d32eebf9738433a42859ecaa83f2c1c272fca47c1ee00a96a26"} Dec 03 06:59:55 crc kubenswrapper[4818]: I1203 06:59:55.559915 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" event={"ID":"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de","Type":"ContainerStarted","Data":"3bc3d91a09348073a9803998c16cbccd9f9f3c00b2788fbafa2b2a8060be7f4d"} Dec 03 06:59:55 crc kubenswrapper[4818]: I1203 06:59:55.583705 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" podStartSLOduration=2.107215303 podStartE2EDuration="2.583687778s" podCreationTimestamp="2025-12-03 06:59:53 +0000 UTC" firstStartedPulling="2025-12-03 06:59:54.594395886 +0000 UTC m=+1952.286004638" lastFinishedPulling="2025-12-03 06:59:55.070868361 +0000 UTC m=+1952.762477113" observedRunningTime="2025-12-03 06:59:55.578585712 +0000 UTC m=+1953.270194464" watchObservedRunningTime="2025-12-03 06:59:55.583687778 +0000 UTC m=+1953.275296520" Dec 03 06:59:55 crc kubenswrapper[4818]: I1203 06:59:55.765449 4818 scope.go:117] "RemoveContainer" containerID="ffbb9c6e2bf23ddfd97de203be0f20b6bd929349b8383901240c6a2c7f15a828" Dec 03 07:00:00 crc kubenswrapper[4818]: I1203 07:00:00.135919 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb"] Dec 03 07:00:00 crc kubenswrapper[4818]: I1203 07:00:00.138806 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb" Dec 03 07:00:00 crc kubenswrapper[4818]: I1203 07:00:00.141354 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 07:00:00 crc kubenswrapper[4818]: I1203 07:00:00.145858 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb"] Dec 03 07:00:00 crc kubenswrapper[4818]: I1203 07:00:00.147409 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 07:00:00 crc kubenswrapper[4818]: I1203 07:00:00.164378 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whkkj\" (UniqueName: \"kubernetes.io/projected/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-kube-api-access-whkkj\") pod \"collect-profiles-29412420-2wcsb\" (UID: \"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb" Dec 03 07:00:00 crc kubenswrapper[4818]: I1203 07:00:00.164513 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-config-volume\") pod \"collect-profiles-29412420-2wcsb\" (UID: \"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb" Dec 03 07:00:00 crc kubenswrapper[4818]: I1203 07:00:00.164921 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-secret-volume\") pod \"collect-profiles-29412420-2wcsb\" (UID: \"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb" Dec 03 07:00:00 crc kubenswrapper[4818]: I1203 07:00:00.265713 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-secret-volume\") pod \"collect-profiles-29412420-2wcsb\" (UID: \"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb" Dec 03 07:00:00 crc kubenswrapper[4818]: I1203 07:00:00.265956 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whkkj\" (UniqueName: \"kubernetes.io/projected/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-kube-api-access-whkkj\") pod \"collect-profiles-29412420-2wcsb\" (UID: \"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb" Dec 03 07:00:00 crc kubenswrapper[4818]: I1203 07:00:00.265988 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-config-volume\") pod \"collect-profiles-29412420-2wcsb\" (UID: \"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb" Dec 03 07:00:00 crc kubenswrapper[4818]: I1203 07:00:00.267200 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-config-volume\") pod \"collect-profiles-29412420-2wcsb\" (UID: \"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb" Dec 03 07:00:00 crc kubenswrapper[4818]: I1203 07:00:00.272146 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-secret-volume\") pod \"collect-profiles-29412420-2wcsb\" (UID: \"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb" Dec 03 07:00:00 crc kubenswrapper[4818]: I1203 07:00:00.292138 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whkkj\" (UniqueName: \"kubernetes.io/projected/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-kube-api-access-whkkj\") pod \"collect-profiles-29412420-2wcsb\" (UID: \"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb" Dec 03 07:00:00 crc kubenswrapper[4818]: I1203 07:00:00.463160 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb" Dec 03 07:00:00 crc kubenswrapper[4818]: I1203 07:00:00.966088 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb"] Dec 03 07:00:01 crc kubenswrapper[4818]: I1203 07:00:01.619133 4818 generic.go:334] "Generic (PLEG): container finished" podID="1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4" containerID="bc78b50a143f1e50878ee867bd6b2dad821660a07ef38d1c2eadd9ce00a2bcfb" exitCode=0 Dec 03 07:00:01 crc kubenswrapper[4818]: I1203 07:00:01.619375 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb" event={"ID":"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4","Type":"ContainerDied","Data":"bc78b50a143f1e50878ee867bd6b2dad821660a07ef38d1c2eadd9ce00a2bcfb"} Dec 03 07:00:01 crc kubenswrapper[4818]: I1203 07:00:01.619422 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb" event={"ID":"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4","Type":"ContainerStarted","Data":"ea32a712cc90627957e4e35b65f4069a3e8eddf347b0dcd1eaa48252c535b0cf"} Dec 03 07:00:03 crc kubenswrapper[4818]: I1203 07:00:03.055078 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb" Dec 03 07:00:03 crc kubenswrapper[4818]: I1203 07:00:03.219993 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whkkj\" (UniqueName: \"kubernetes.io/projected/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-kube-api-access-whkkj\") pod \"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4\" (UID: \"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4\") " Dec 03 07:00:03 crc kubenswrapper[4818]: I1203 07:00:03.220162 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-config-volume\") pod \"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4\" (UID: \"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4\") " Dec 03 07:00:03 crc kubenswrapper[4818]: I1203 07:00:03.220212 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-secret-volume\") pod \"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4\" (UID: \"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4\") " Dec 03 07:00:03 crc kubenswrapper[4818]: I1203 07:00:03.221152 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-config-volume" (OuterVolumeSpecName: "config-volume") pod "1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4" (UID: "1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:00:03 crc kubenswrapper[4818]: I1203 07:00:03.228031 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4" (UID: "1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:00:03 crc kubenswrapper[4818]: I1203 07:00:03.231245 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-kube-api-access-whkkj" (OuterVolumeSpecName: "kube-api-access-whkkj") pod "1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4" (UID: "1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4"). InnerVolumeSpecName "kube-api-access-whkkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:00:03 crc kubenswrapper[4818]: I1203 07:00:03.322787 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whkkj\" (UniqueName: \"kubernetes.io/projected/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-kube-api-access-whkkj\") on node \"crc\" DevicePath \"\"" Dec 03 07:00:03 crc kubenswrapper[4818]: I1203 07:00:03.322877 4818 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 07:00:03 crc kubenswrapper[4818]: I1203 07:00:03.322897 4818 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 07:00:03 crc kubenswrapper[4818]: I1203 07:00:03.643124 4818 generic.go:334] "Generic (PLEG): container finished" podID="b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de" containerID="18213a6eecd26d32eebf9738433a42859ecaa83f2c1c272fca47c1ee00a96a26" exitCode=0 Dec 03 07:00:03 crc kubenswrapper[4818]: I1203 07:00:03.643214 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" event={"ID":"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de","Type":"ContainerDied","Data":"18213a6eecd26d32eebf9738433a42859ecaa83f2c1c272fca47c1ee00a96a26"} Dec 03 07:00:03 crc kubenswrapper[4818]: I1203 07:00:03.646453 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb" event={"ID":"1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4","Type":"ContainerDied","Data":"ea32a712cc90627957e4e35b65f4069a3e8eddf347b0dcd1eaa48252c535b0cf"} Dec 03 07:00:03 crc kubenswrapper[4818]: I1203 07:00:03.646491 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea32a712cc90627957e4e35b65f4069a3e8eddf347b0dcd1eaa48252c535b0cf" Dec 03 07:00:03 crc kubenswrapper[4818]: I1203 07:00:03.646532 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412420-2wcsb" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.200789 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.384161 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jm6kc\" (UniqueName: \"kubernetes.io/projected/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-kube-api-access-jm6kc\") pod \"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de\" (UID: \"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de\") " Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.384462 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-inventory-0\") pod \"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de\" (UID: \"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de\") " Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.384544 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-ssh-key-openstack-edpm-ipam\") pod \"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de\" (UID: \"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de\") " Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.389688 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-kube-api-access-jm6kc" (OuterVolumeSpecName: "kube-api-access-jm6kc") pod "b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de" (UID: "b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de"). InnerVolumeSpecName "kube-api-access-jm6kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.416496 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de" (UID: "b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.417936 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de" (UID: "b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.487297 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jm6kc\" (UniqueName: \"kubernetes.io/projected/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-kube-api-access-jm6kc\") on node \"crc\" DevicePath \"\"" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.487334 4818 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.487345 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.664499 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" event={"ID":"b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de","Type":"ContainerDied","Data":"3bc3d91a09348073a9803998c16cbccd9f9f3c00b2788fbafa2b2a8060be7f4d"} Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.664537 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bc3d91a09348073a9803998c16cbccd9f9f3c00b2788fbafa2b2a8060be7f4d" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.664787 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-lbkfx" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.751669 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l"] Dec 03 07:00:05 crc kubenswrapper[4818]: E1203 07:00:05.752096 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de" containerName="ssh-known-hosts-edpm-deployment" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.752117 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de" containerName="ssh-known-hosts-edpm-deployment" Dec 03 07:00:05 crc kubenswrapper[4818]: E1203 07:00:05.752138 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4" containerName="collect-profiles" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.752144 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4" containerName="collect-profiles" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.752329 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dcaa8ab-b22b-434f-bcd7-5efa4705bdc4" containerName="collect-profiles" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.752365 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de" containerName="ssh-known-hosts-edpm-deployment" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.755631 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.761993 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2bpsv" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.761996 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.762131 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.762457 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.765248 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l"] Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.793903 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90add555-879b-4da5-a350-40ce938e9956-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d689l\" (UID: \"90add555-879b-4da5-a350-40ce938e9956\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.794129 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90add555-879b-4da5-a350-40ce938e9956-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d689l\" (UID: \"90add555-879b-4da5-a350-40ce938e9956\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.794179 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjhlh\" (UniqueName: \"kubernetes.io/projected/90add555-879b-4da5-a350-40ce938e9956-kube-api-access-pjhlh\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d689l\" (UID: \"90add555-879b-4da5-a350-40ce938e9956\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.894831 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90add555-879b-4da5-a350-40ce938e9956-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d689l\" (UID: \"90add555-879b-4da5-a350-40ce938e9956\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.895002 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90add555-879b-4da5-a350-40ce938e9956-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d689l\" (UID: \"90add555-879b-4da5-a350-40ce938e9956\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.895078 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjhlh\" (UniqueName: \"kubernetes.io/projected/90add555-879b-4da5-a350-40ce938e9956-kube-api-access-pjhlh\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d689l\" (UID: \"90add555-879b-4da5-a350-40ce938e9956\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.900169 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90add555-879b-4da5-a350-40ce938e9956-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d689l\" (UID: \"90add555-879b-4da5-a350-40ce938e9956\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.900483 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90add555-879b-4da5-a350-40ce938e9956-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d689l\" (UID: \"90add555-879b-4da5-a350-40ce938e9956\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" Dec 03 07:00:05 crc kubenswrapper[4818]: I1203 07:00:05.918721 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjhlh\" (UniqueName: \"kubernetes.io/projected/90add555-879b-4da5-a350-40ce938e9956-kube-api-access-pjhlh\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-d689l\" (UID: \"90add555-879b-4da5-a350-40ce938e9956\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" Dec 03 07:00:06 crc kubenswrapper[4818]: I1203 07:00:06.080301 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" Dec 03 07:00:06 crc kubenswrapper[4818]: I1203 07:00:06.751444 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l"] Dec 03 07:00:07 crc kubenswrapper[4818]: I1203 07:00:07.726646 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" event={"ID":"90add555-879b-4da5-a350-40ce938e9956","Type":"ContainerStarted","Data":"e99534bd3b00c0d2aa0f724bcd93ccb1a2a851e96f604067fafe4d16418a1236"} Dec 03 07:00:07 crc kubenswrapper[4818]: I1203 07:00:07.726700 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" event={"ID":"90add555-879b-4da5-a350-40ce938e9956","Type":"ContainerStarted","Data":"59c03a49a3a223429a63fce460e363ce8c505e16c667122d35cf529d43ed1090"} Dec 03 07:00:07 crc kubenswrapper[4818]: I1203 07:00:07.755132 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" podStartSLOduration=2.303491813 podStartE2EDuration="2.755100391s" podCreationTimestamp="2025-12-03 07:00:05 +0000 UTC" firstStartedPulling="2025-12-03 07:00:06.708832306 +0000 UTC m=+1964.400441058" lastFinishedPulling="2025-12-03 07:00:07.160440874 +0000 UTC m=+1964.852049636" observedRunningTime="2025-12-03 07:00:07.753084881 +0000 UTC m=+1965.444693683" watchObservedRunningTime="2025-12-03 07:00:07.755100391 +0000 UTC m=+1965.446709183" Dec 03 07:00:16 crc kubenswrapper[4818]: I1203 07:00:16.873526 4818 generic.go:334] "Generic (PLEG): container finished" podID="90add555-879b-4da5-a350-40ce938e9956" containerID="e99534bd3b00c0d2aa0f724bcd93ccb1a2a851e96f604067fafe4d16418a1236" exitCode=0 Dec 03 07:00:16 crc kubenswrapper[4818]: I1203 07:00:16.873654 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" event={"ID":"90add555-879b-4da5-a350-40ce938e9956","Type":"ContainerDied","Data":"e99534bd3b00c0d2aa0f724bcd93ccb1a2a851e96f604067fafe4d16418a1236"} Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.373879 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.420744 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90add555-879b-4da5-a350-40ce938e9956-ssh-key\") pod \"90add555-879b-4da5-a350-40ce938e9956\" (UID: \"90add555-879b-4da5-a350-40ce938e9956\") " Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.466835 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90add555-879b-4da5-a350-40ce938e9956-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "90add555-879b-4da5-a350-40ce938e9956" (UID: "90add555-879b-4da5-a350-40ce938e9956"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.521687 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjhlh\" (UniqueName: \"kubernetes.io/projected/90add555-879b-4da5-a350-40ce938e9956-kube-api-access-pjhlh\") pod \"90add555-879b-4da5-a350-40ce938e9956\" (UID: \"90add555-879b-4da5-a350-40ce938e9956\") " Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.521915 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90add555-879b-4da5-a350-40ce938e9956-inventory\") pod \"90add555-879b-4da5-a350-40ce938e9956\" (UID: \"90add555-879b-4da5-a350-40ce938e9956\") " Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.522426 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90add555-879b-4da5-a350-40ce938e9956-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.525854 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90add555-879b-4da5-a350-40ce938e9956-kube-api-access-pjhlh" (OuterVolumeSpecName: "kube-api-access-pjhlh") pod "90add555-879b-4da5-a350-40ce938e9956" (UID: "90add555-879b-4da5-a350-40ce938e9956"). InnerVolumeSpecName "kube-api-access-pjhlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.547408 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90add555-879b-4da5-a350-40ce938e9956-inventory" (OuterVolumeSpecName: "inventory") pod "90add555-879b-4da5-a350-40ce938e9956" (UID: "90add555-879b-4da5-a350-40ce938e9956"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.624614 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjhlh\" (UniqueName: \"kubernetes.io/projected/90add555-879b-4da5-a350-40ce938e9956-kube-api-access-pjhlh\") on node \"crc\" DevicePath \"\"" Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.624648 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90add555-879b-4da5-a350-40ce938e9956-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.892496 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" event={"ID":"90add555-879b-4da5-a350-40ce938e9956","Type":"ContainerDied","Data":"59c03a49a3a223429a63fce460e363ce8c505e16c667122d35cf529d43ed1090"} Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.892533 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59c03a49a3a223429a63fce460e363ce8c505e16c667122d35cf529d43ed1090" Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.892568 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-d689l" Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.965434 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26"] Dec 03 07:00:18 crc kubenswrapper[4818]: E1203 07:00:18.966309 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90add555-879b-4da5-a350-40ce938e9956" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.966336 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="90add555-879b-4da5-a350-40ce938e9956" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.966591 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="90add555-879b-4da5-a350-40ce938e9956" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.967310 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26" Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.969218 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.969953 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.970293 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.970462 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2bpsv" Dec 03 07:00:18 crc kubenswrapper[4818]: I1203 07:00:18.976295 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26"] Dec 03 07:00:19 crc kubenswrapper[4818]: I1203 07:00:19.031176 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26\" (UID: \"3bf68a0c-4bbc-4f9e-b3be-503d857ead66\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26" Dec 03 07:00:19 crc kubenswrapper[4818]: I1203 07:00:19.031250 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwtcf\" (UniqueName: \"kubernetes.io/projected/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-kube-api-access-jwtcf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26\" (UID: \"3bf68a0c-4bbc-4f9e-b3be-503d857ead66\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26" Dec 03 07:00:19 crc kubenswrapper[4818]: I1203 07:00:19.031312 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26\" (UID: \"3bf68a0c-4bbc-4f9e-b3be-503d857ead66\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26" Dec 03 07:00:19 crc kubenswrapper[4818]: I1203 07:00:19.132345 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26\" (UID: \"3bf68a0c-4bbc-4f9e-b3be-503d857ead66\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26" Dec 03 07:00:19 crc kubenswrapper[4818]: I1203 07:00:19.132591 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwtcf\" (UniqueName: \"kubernetes.io/projected/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-kube-api-access-jwtcf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26\" (UID: \"3bf68a0c-4bbc-4f9e-b3be-503d857ead66\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26" Dec 03 07:00:19 crc kubenswrapper[4818]: I1203 07:00:19.132724 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26\" (UID: \"3bf68a0c-4bbc-4f9e-b3be-503d857ead66\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26" Dec 03 07:00:19 crc kubenswrapper[4818]: I1203 07:00:19.137678 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26\" (UID: \"3bf68a0c-4bbc-4f9e-b3be-503d857ead66\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26" Dec 03 07:00:19 crc kubenswrapper[4818]: I1203 07:00:19.137825 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26\" (UID: \"3bf68a0c-4bbc-4f9e-b3be-503d857ead66\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26" Dec 03 07:00:19 crc kubenswrapper[4818]: I1203 07:00:19.151050 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwtcf\" (UniqueName: \"kubernetes.io/projected/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-kube-api-access-jwtcf\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26\" (UID: \"3bf68a0c-4bbc-4f9e-b3be-503d857ead66\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26" Dec 03 07:00:19 crc kubenswrapper[4818]: I1203 07:00:19.304523 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26" Dec 03 07:00:19 crc kubenswrapper[4818]: I1203 07:00:19.857973 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26"] Dec 03 07:00:19 crc kubenswrapper[4818]: W1203 07:00:19.862715 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3bf68a0c_4bbc_4f9e_b3be_503d857ead66.slice/crio-d681ed4ebafdf4169062385b0d7676ad2549b5519ebbb887bde1e58094059ac6 WatchSource:0}: Error finding container d681ed4ebafdf4169062385b0d7676ad2549b5519ebbb887bde1e58094059ac6: Status 404 returned error can't find the container with id d681ed4ebafdf4169062385b0d7676ad2549b5519ebbb887bde1e58094059ac6 Dec 03 07:00:19 crc kubenswrapper[4818]: I1203 07:00:19.902868 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26" event={"ID":"3bf68a0c-4bbc-4f9e-b3be-503d857ead66","Type":"ContainerStarted","Data":"d681ed4ebafdf4169062385b0d7676ad2549b5519ebbb887bde1e58094059ac6"} Dec 03 07:00:20 crc kubenswrapper[4818]: I1203 07:00:20.911908 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26" event={"ID":"3bf68a0c-4bbc-4f9e-b3be-503d857ead66","Type":"ContainerStarted","Data":"fac1600f0557d193cbd0278e986417d892cc06281e23028fa4b07e93491da41c"} Dec 03 07:00:31 crc kubenswrapper[4818]: I1203 07:00:31.007929 4818 generic.go:334] "Generic (PLEG): container finished" podID="3bf68a0c-4bbc-4f9e-b3be-503d857ead66" containerID="fac1600f0557d193cbd0278e986417d892cc06281e23028fa4b07e93491da41c" exitCode=0 Dec 03 07:00:31 crc kubenswrapper[4818]: I1203 07:00:31.008036 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26" event={"ID":"3bf68a0c-4bbc-4f9e-b3be-503d857ead66","Type":"ContainerDied","Data":"fac1600f0557d193cbd0278e986417d892cc06281e23028fa4b07e93491da41c"} Dec 03 07:00:32 crc kubenswrapper[4818]: I1203 07:00:32.453855 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26" Dec 03 07:00:32 crc kubenswrapper[4818]: I1203 07:00:32.612636 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwtcf\" (UniqueName: \"kubernetes.io/projected/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-kube-api-access-jwtcf\") pod \"3bf68a0c-4bbc-4f9e-b3be-503d857ead66\" (UID: \"3bf68a0c-4bbc-4f9e-b3be-503d857ead66\") " Dec 03 07:00:32 crc kubenswrapper[4818]: I1203 07:00:32.612801 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-ssh-key\") pod \"3bf68a0c-4bbc-4f9e-b3be-503d857ead66\" (UID: \"3bf68a0c-4bbc-4f9e-b3be-503d857ead66\") " Dec 03 07:00:32 crc kubenswrapper[4818]: I1203 07:00:32.612948 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-inventory\") pod \"3bf68a0c-4bbc-4f9e-b3be-503d857ead66\" (UID: \"3bf68a0c-4bbc-4f9e-b3be-503d857ead66\") " Dec 03 07:00:32 crc kubenswrapper[4818]: I1203 07:00:32.619130 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-kube-api-access-jwtcf" (OuterVolumeSpecName: "kube-api-access-jwtcf") pod "3bf68a0c-4bbc-4f9e-b3be-503d857ead66" (UID: "3bf68a0c-4bbc-4f9e-b3be-503d857ead66"). InnerVolumeSpecName "kube-api-access-jwtcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:00:32 crc kubenswrapper[4818]: I1203 07:00:32.650084 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-inventory" (OuterVolumeSpecName: "inventory") pod "3bf68a0c-4bbc-4f9e-b3be-503d857ead66" (UID: "3bf68a0c-4bbc-4f9e-b3be-503d857ead66"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:00:32 crc kubenswrapper[4818]: I1203 07:00:32.657055 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3bf68a0c-4bbc-4f9e-b3be-503d857ead66" (UID: "3bf68a0c-4bbc-4f9e-b3be-503d857ead66"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:00:32 crc kubenswrapper[4818]: I1203 07:00:32.715411 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 07:00:32 crc kubenswrapper[4818]: I1203 07:00:32.715451 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 07:00:32 crc kubenswrapper[4818]: I1203 07:00:32.715472 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwtcf\" (UniqueName: \"kubernetes.io/projected/3bf68a0c-4bbc-4f9e-b3be-503d857ead66-kube-api-access-jwtcf\") on node \"crc\" DevicePath \"\"" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.043260 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26" event={"ID":"3bf68a0c-4bbc-4f9e-b3be-503d857ead66","Type":"ContainerDied","Data":"d681ed4ebafdf4169062385b0d7676ad2549b5519ebbb887bde1e58094059ac6"} Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.043323 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.043334 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d681ed4ebafdf4169062385b0d7676ad2549b5519ebbb887bde1e58094059ac6" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.153126 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5"] Dec 03 07:00:33 crc kubenswrapper[4818]: E1203 07:00:33.153986 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bf68a0c-4bbc-4f9e-b3be-503d857ead66" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.154012 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bf68a0c-4bbc-4f9e-b3be-503d857ead66" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.154366 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bf68a0c-4bbc-4f9e-b3be-503d857ead66" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.155298 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.157996 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.159447 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.159607 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2bpsv" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.163050 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.163265 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.164807 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.165387 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.165571 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.168217 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5"] Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.325985 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.326070 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.326108 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.326161 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.326186 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.326209 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.326299 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.326357 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.326397 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.326596 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.326646 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.326671 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwkw5\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-kube-api-access-dwkw5\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.326759 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.326903 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.429224 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.429293 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.429354 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.429386 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.429409 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.429433 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.429456 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.429498 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.429539 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.429571 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.429622 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.429646 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.429667 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwkw5\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-kube-api-access-dwkw5\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.429711 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.436481 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.436542 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.437444 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.437740 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.437777 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.438399 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.439270 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.439576 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.440549 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.444946 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.445003 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.449630 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.449739 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.465973 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwkw5\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-kube-api-access-dwkw5\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z89w5\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:33 crc kubenswrapper[4818]: I1203 07:00:33.475963 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:00:34 crc kubenswrapper[4818]: I1203 07:00:34.020722 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5"] Dec 03 07:00:34 crc kubenswrapper[4818]: I1203 07:00:34.053532 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" event={"ID":"1808af84-64d0-4b2a-a5bf-32653bb3c948","Type":"ContainerStarted","Data":"b4ab4f018ce8d69b501532c8874c8abd45edc173a049d29d2f35e6f3d9f7115c"} Dec 03 07:00:35 crc kubenswrapper[4818]: I1203 07:00:35.064877 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" event={"ID":"1808af84-64d0-4b2a-a5bf-32653bb3c948","Type":"ContainerStarted","Data":"39b1aa18a301738ccdc544ae24db889b5b1a5ad120e0d1fdcf6545ea9355b689"} Dec 03 07:00:35 crc kubenswrapper[4818]: I1203 07:00:35.088205 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" podStartSLOduration=1.429881504 podStartE2EDuration="2.088165051s" podCreationTimestamp="2025-12-03 07:00:33 +0000 UTC" firstStartedPulling="2025-12-03 07:00:34.024177715 +0000 UTC m=+1991.715786467" lastFinishedPulling="2025-12-03 07:00:34.682461262 +0000 UTC m=+1992.374070014" observedRunningTime="2025-12-03 07:00:35.085596827 +0000 UTC m=+1992.777205579" watchObservedRunningTime="2025-12-03 07:00:35.088165051 +0000 UTC m=+1992.779773803" Dec 03 07:00:43 crc kubenswrapper[4818]: I1203 07:00:43.302074 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:00:43 crc kubenswrapper[4818]: I1203 07:00:43.302667 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:01:00 crc kubenswrapper[4818]: I1203 07:01:00.152382 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29412421-4n5xh"] Dec 03 07:01:00 crc kubenswrapper[4818]: I1203 07:01:00.154947 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412421-4n5xh" Dec 03 07:01:00 crc kubenswrapper[4818]: I1203 07:01:00.167529 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29412421-4n5xh"] Dec 03 07:01:00 crc kubenswrapper[4818]: I1203 07:01:00.329970 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-combined-ca-bundle\") pod \"keystone-cron-29412421-4n5xh\" (UID: \"658c06c1-51a3-467d-8de0-2205f7c22cbc\") " pod="openstack/keystone-cron-29412421-4n5xh" Dec 03 07:01:00 crc kubenswrapper[4818]: I1203 07:01:00.330074 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-config-data\") pod \"keystone-cron-29412421-4n5xh\" (UID: \"658c06c1-51a3-467d-8de0-2205f7c22cbc\") " pod="openstack/keystone-cron-29412421-4n5xh" Dec 03 07:01:00 crc kubenswrapper[4818]: I1203 07:01:00.330169 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpqdn\" (UniqueName: \"kubernetes.io/projected/658c06c1-51a3-467d-8de0-2205f7c22cbc-kube-api-access-fpqdn\") pod \"keystone-cron-29412421-4n5xh\" (UID: \"658c06c1-51a3-467d-8de0-2205f7c22cbc\") " pod="openstack/keystone-cron-29412421-4n5xh" Dec 03 07:01:00 crc kubenswrapper[4818]: I1203 07:01:00.330363 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-fernet-keys\") pod \"keystone-cron-29412421-4n5xh\" (UID: \"658c06c1-51a3-467d-8de0-2205f7c22cbc\") " pod="openstack/keystone-cron-29412421-4n5xh" Dec 03 07:01:00 crc kubenswrapper[4818]: I1203 07:01:00.432424 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-fernet-keys\") pod \"keystone-cron-29412421-4n5xh\" (UID: \"658c06c1-51a3-467d-8de0-2205f7c22cbc\") " pod="openstack/keystone-cron-29412421-4n5xh" Dec 03 07:01:00 crc kubenswrapper[4818]: I1203 07:01:00.432759 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-combined-ca-bundle\") pod \"keystone-cron-29412421-4n5xh\" (UID: \"658c06c1-51a3-467d-8de0-2205f7c22cbc\") " pod="openstack/keystone-cron-29412421-4n5xh" Dec 03 07:01:00 crc kubenswrapper[4818]: I1203 07:01:00.432944 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-config-data\") pod \"keystone-cron-29412421-4n5xh\" (UID: \"658c06c1-51a3-467d-8de0-2205f7c22cbc\") " pod="openstack/keystone-cron-29412421-4n5xh" Dec 03 07:01:00 crc kubenswrapper[4818]: I1203 07:01:00.433115 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpqdn\" (UniqueName: \"kubernetes.io/projected/658c06c1-51a3-467d-8de0-2205f7c22cbc-kube-api-access-fpqdn\") pod \"keystone-cron-29412421-4n5xh\" (UID: \"658c06c1-51a3-467d-8de0-2205f7c22cbc\") " pod="openstack/keystone-cron-29412421-4n5xh" Dec 03 07:01:00 crc kubenswrapper[4818]: I1203 07:01:00.441168 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-config-data\") pod \"keystone-cron-29412421-4n5xh\" (UID: \"658c06c1-51a3-467d-8de0-2205f7c22cbc\") " pod="openstack/keystone-cron-29412421-4n5xh" Dec 03 07:01:00 crc kubenswrapper[4818]: I1203 07:01:00.441393 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-fernet-keys\") pod \"keystone-cron-29412421-4n5xh\" (UID: \"658c06c1-51a3-467d-8de0-2205f7c22cbc\") " pod="openstack/keystone-cron-29412421-4n5xh" Dec 03 07:01:00 crc kubenswrapper[4818]: I1203 07:01:00.444837 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-combined-ca-bundle\") pod \"keystone-cron-29412421-4n5xh\" (UID: \"658c06c1-51a3-467d-8de0-2205f7c22cbc\") " pod="openstack/keystone-cron-29412421-4n5xh" Dec 03 07:01:00 crc kubenswrapper[4818]: I1203 07:01:00.451006 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpqdn\" (UniqueName: \"kubernetes.io/projected/658c06c1-51a3-467d-8de0-2205f7c22cbc-kube-api-access-fpqdn\") pod \"keystone-cron-29412421-4n5xh\" (UID: \"658c06c1-51a3-467d-8de0-2205f7c22cbc\") " pod="openstack/keystone-cron-29412421-4n5xh" Dec 03 07:01:00 crc kubenswrapper[4818]: I1203 07:01:00.479364 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412421-4n5xh" Dec 03 07:01:00 crc kubenswrapper[4818]: I1203 07:01:00.996538 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29412421-4n5xh"] Dec 03 07:01:01 crc kubenswrapper[4818]: I1203 07:01:01.306635 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412421-4n5xh" event={"ID":"658c06c1-51a3-467d-8de0-2205f7c22cbc","Type":"ContainerStarted","Data":"e2e0f0a80fc2809832acdd4be915900b08d7968cf8810100ba5790721f91a229"} Dec 03 07:01:01 crc kubenswrapper[4818]: I1203 07:01:01.306680 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412421-4n5xh" event={"ID":"658c06c1-51a3-467d-8de0-2205f7c22cbc","Type":"ContainerStarted","Data":"61b3afdf3d218bb839903b729747ac45dbd53f8ffd7a8bf4ebbe448a54760601"} Dec 03 07:01:01 crc kubenswrapper[4818]: I1203 07:01:01.330963 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29412421-4n5xh" podStartSLOduration=1.3309465600000001 podStartE2EDuration="1.33094656s" podCreationTimestamp="2025-12-03 07:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:01:01.320316627 +0000 UTC m=+2019.011925379" watchObservedRunningTime="2025-12-03 07:01:01.33094656 +0000 UTC m=+2019.022555312" Dec 03 07:01:03 crc kubenswrapper[4818]: I1203 07:01:03.329334 4818 generic.go:334] "Generic (PLEG): container finished" podID="658c06c1-51a3-467d-8de0-2205f7c22cbc" containerID="e2e0f0a80fc2809832acdd4be915900b08d7968cf8810100ba5790721f91a229" exitCode=0 Dec 03 07:01:03 crc kubenswrapper[4818]: I1203 07:01:03.329437 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412421-4n5xh" event={"ID":"658c06c1-51a3-467d-8de0-2205f7c22cbc","Type":"ContainerDied","Data":"e2e0f0a80fc2809832acdd4be915900b08d7968cf8810100ba5790721f91a229"} Dec 03 07:01:04 crc kubenswrapper[4818]: I1203 07:01:04.744774 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412421-4n5xh" Dec 03 07:01:04 crc kubenswrapper[4818]: I1203 07:01:04.848847 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpqdn\" (UniqueName: \"kubernetes.io/projected/658c06c1-51a3-467d-8de0-2205f7c22cbc-kube-api-access-fpqdn\") pod \"658c06c1-51a3-467d-8de0-2205f7c22cbc\" (UID: \"658c06c1-51a3-467d-8de0-2205f7c22cbc\") " Dec 03 07:01:04 crc kubenswrapper[4818]: I1203 07:01:04.849096 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-config-data\") pod \"658c06c1-51a3-467d-8de0-2205f7c22cbc\" (UID: \"658c06c1-51a3-467d-8de0-2205f7c22cbc\") " Dec 03 07:01:04 crc kubenswrapper[4818]: I1203 07:01:04.849183 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-combined-ca-bundle\") pod \"658c06c1-51a3-467d-8de0-2205f7c22cbc\" (UID: \"658c06c1-51a3-467d-8de0-2205f7c22cbc\") " Dec 03 07:01:04 crc kubenswrapper[4818]: I1203 07:01:04.849260 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-fernet-keys\") pod \"658c06c1-51a3-467d-8de0-2205f7c22cbc\" (UID: \"658c06c1-51a3-467d-8de0-2205f7c22cbc\") " Dec 03 07:01:04 crc kubenswrapper[4818]: I1203 07:01:04.855061 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/658c06c1-51a3-467d-8de0-2205f7c22cbc-kube-api-access-fpqdn" (OuterVolumeSpecName: "kube-api-access-fpqdn") pod "658c06c1-51a3-467d-8de0-2205f7c22cbc" (UID: "658c06c1-51a3-467d-8de0-2205f7c22cbc"). InnerVolumeSpecName "kube-api-access-fpqdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:01:04 crc kubenswrapper[4818]: I1203 07:01:04.864843 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "658c06c1-51a3-467d-8de0-2205f7c22cbc" (UID: "658c06c1-51a3-467d-8de0-2205f7c22cbc"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:01:04 crc kubenswrapper[4818]: I1203 07:01:04.891640 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "658c06c1-51a3-467d-8de0-2205f7c22cbc" (UID: "658c06c1-51a3-467d-8de0-2205f7c22cbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:01:04 crc kubenswrapper[4818]: I1203 07:01:04.905508 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-config-data" (OuterVolumeSpecName: "config-data") pod "658c06c1-51a3-467d-8de0-2205f7c22cbc" (UID: "658c06c1-51a3-467d-8de0-2205f7c22cbc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:01:04 crc kubenswrapper[4818]: I1203 07:01:04.955545 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:04 crc kubenswrapper[4818]: I1203 07:01:04.956122 4818 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:04 crc kubenswrapper[4818]: I1203 07:01:04.956157 4818 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/658c06c1-51a3-467d-8de0-2205f7c22cbc-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:04 crc kubenswrapper[4818]: I1203 07:01:04.956184 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpqdn\" (UniqueName: \"kubernetes.io/projected/658c06c1-51a3-467d-8de0-2205f7c22cbc-kube-api-access-fpqdn\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:05 crc kubenswrapper[4818]: I1203 07:01:05.349371 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412421-4n5xh" Dec 03 07:01:05 crc kubenswrapper[4818]: I1203 07:01:05.349409 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412421-4n5xh" event={"ID":"658c06c1-51a3-467d-8de0-2205f7c22cbc","Type":"ContainerDied","Data":"61b3afdf3d218bb839903b729747ac45dbd53f8ffd7a8bf4ebbe448a54760601"} Dec 03 07:01:05 crc kubenswrapper[4818]: I1203 07:01:05.349570 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61b3afdf3d218bb839903b729747ac45dbd53f8ffd7a8bf4ebbe448a54760601" Dec 03 07:01:13 crc kubenswrapper[4818]: I1203 07:01:13.375051 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:01:13 crc kubenswrapper[4818]: I1203 07:01:13.375641 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:01:14 crc kubenswrapper[4818]: I1203 07:01:14.428422 4818 generic.go:334] "Generic (PLEG): container finished" podID="1808af84-64d0-4b2a-a5bf-32653bb3c948" containerID="39b1aa18a301738ccdc544ae24db889b5b1a5ad120e0d1fdcf6545ea9355b689" exitCode=0 Dec 03 07:01:14 crc kubenswrapper[4818]: I1203 07:01:14.428500 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" event={"ID":"1808af84-64d0-4b2a-a5bf-32653bb3c948","Type":"ContainerDied","Data":"39b1aa18a301738ccdc544ae24db889b5b1a5ad120e0d1fdcf6545ea9355b689"} Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.835250 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.970170 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-inventory\") pod \"1808af84-64d0-4b2a-a5bf-32653bb3c948\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.970254 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-repo-setup-combined-ca-bundle\") pod \"1808af84-64d0-4b2a-a5bf-32653bb3c948\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.970277 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"1808af84-64d0-4b2a-a5bf-32653bb3c948\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.970321 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-neutron-metadata-combined-ca-bundle\") pod \"1808af84-64d0-4b2a-a5bf-32653bb3c948\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.970350 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-nova-combined-ca-bundle\") pod \"1808af84-64d0-4b2a-a5bf-32653bb3c948\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.970375 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-ovn-combined-ca-bundle\") pod \"1808af84-64d0-4b2a-a5bf-32653bb3c948\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.970427 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"1808af84-64d0-4b2a-a5bf-32653bb3c948\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.970447 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwkw5\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-kube-api-access-dwkw5\") pod \"1808af84-64d0-4b2a-a5bf-32653bb3c948\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.970461 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"1808af84-64d0-4b2a-a5bf-32653bb3c948\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.970520 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-bootstrap-combined-ca-bundle\") pod \"1808af84-64d0-4b2a-a5bf-32653bb3c948\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.970571 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-ovn-default-certs-0\") pod \"1808af84-64d0-4b2a-a5bf-32653bb3c948\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.970588 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-libvirt-combined-ca-bundle\") pod \"1808af84-64d0-4b2a-a5bf-32653bb3c948\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.970633 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-telemetry-combined-ca-bundle\") pod \"1808af84-64d0-4b2a-a5bf-32653bb3c948\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.970700 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-ssh-key\") pod \"1808af84-64d0-4b2a-a5bf-32653bb3c948\" (UID: \"1808af84-64d0-4b2a-a5bf-32653bb3c948\") " Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.977322 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "1808af84-64d0-4b2a-a5bf-32653bb3c948" (UID: "1808af84-64d0-4b2a-a5bf-32653bb3c948"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.978805 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "1808af84-64d0-4b2a-a5bf-32653bb3c948" (UID: "1808af84-64d0-4b2a-a5bf-32653bb3c948"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.978852 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-kube-api-access-dwkw5" (OuterVolumeSpecName: "kube-api-access-dwkw5") pod "1808af84-64d0-4b2a-a5bf-32653bb3c948" (UID: "1808af84-64d0-4b2a-a5bf-32653bb3c948"). InnerVolumeSpecName "kube-api-access-dwkw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.978931 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "1808af84-64d0-4b2a-a5bf-32653bb3c948" (UID: "1808af84-64d0-4b2a-a5bf-32653bb3c948"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.979022 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "1808af84-64d0-4b2a-a5bf-32653bb3c948" (UID: "1808af84-64d0-4b2a-a5bf-32653bb3c948"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.979176 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "1808af84-64d0-4b2a-a5bf-32653bb3c948" (UID: "1808af84-64d0-4b2a-a5bf-32653bb3c948"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.979705 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "1808af84-64d0-4b2a-a5bf-32653bb3c948" (UID: "1808af84-64d0-4b2a-a5bf-32653bb3c948"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.980403 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "1808af84-64d0-4b2a-a5bf-32653bb3c948" (UID: "1808af84-64d0-4b2a-a5bf-32653bb3c948"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.980745 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "1808af84-64d0-4b2a-a5bf-32653bb3c948" (UID: "1808af84-64d0-4b2a-a5bf-32653bb3c948"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.981163 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "1808af84-64d0-4b2a-a5bf-32653bb3c948" (UID: "1808af84-64d0-4b2a-a5bf-32653bb3c948"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.981605 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "1808af84-64d0-4b2a-a5bf-32653bb3c948" (UID: "1808af84-64d0-4b2a-a5bf-32653bb3c948"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:01:15 crc kubenswrapper[4818]: I1203 07:01:15.989285 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "1808af84-64d0-4b2a-a5bf-32653bb3c948" (UID: "1808af84-64d0-4b2a-a5bf-32653bb3c948"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.002832 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-inventory" (OuterVolumeSpecName: "inventory") pod "1808af84-64d0-4b2a-a5bf-32653bb3c948" (UID: "1808af84-64d0-4b2a-a5bf-32653bb3c948"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.005669 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1808af84-64d0-4b2a-a5bf-32653bb3c948" (UID: "1808af84-64d0-4b2a-a5bf-32653bb3c948"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.075354 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.075392 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.075405 4818 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.075420 4818 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.075431 4818 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.075441 4818 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.075453 4818 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.075462 4818 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.075471 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwkw5\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-kube-api-access-dwkw5\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.075482 4818 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.075491 4818 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.075500 4818 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1808af84-64d0-4b2a-a5bf-32653bb3c948-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.075509 4818 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.075518 4818 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1808af84-64d0-4b2a-a5bf-32653bb3c948-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.448105 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" event={"ID":"1808af84-64d0-4b2a-a5bf-32653bb3c948","Type":"ContainerDied","Data":"b4ab4f018ce8d69b501532c8874c8abd45edc173a049d29d2f35e6f3d9f7115c"} Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.448159 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4ab4f018ce8d69b501532c8874c8abd45edc173a049d29d2f35e6f3d9f7115c" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.448248 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z89w5" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.542479 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7"] Dec 03 07:01:16 crc kubenswrapper[4818]: E1203 07:01:16.543106 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1808af84-64d0-4b2a-a5bf-32653bb3c948" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.543126 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="1808af84-64d0-4b2a-a5bf-32653bb3c948" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 07:01:16 crc kubenswrapper[4818]: E1203 07:01:16.543140 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="658c06c1-51a3-467d-8de0-2205f7c22cbc" containerName="keystone-cron" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.543146 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="658c06c1-51a3-467d-8de0-2205f7c22cbc" containerName="keystone-cron" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.543337 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="658c06c1-51a3-467d-8de0-2205f7c22cbc" containerName="keystone-cron" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.543362 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="1808af84-64d0-4b2a-a5bf-32653bb3c948" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.544237 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.547023 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.547057 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.547104 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.547257 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2bpsv" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.547357 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.552661 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7"] Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.707009 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmsk6\" (UniqueName: \"kubernetes.io/projected/22cbc8e9-1434-4dcb-8fd3-24def21fe451-kube-api-access-jmsk6\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xxxp7\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.707650 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xxxp7\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.707926 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xxxp7\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.708184 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xxxp7\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.708498 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xxxp7\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.810707 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xxxp7\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.811090 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xxxp7\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.811178 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xxxp7\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.811247 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xxxp7\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.811370 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmsk6\" (UniqueName: \"kubernetes.io/projected/22cbc8e9-1434-4dcb-8fd3-24def21fe451-kube-api-access-jmsk6\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xxxp7\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.812378 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xxxp7\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.814858 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xxxp7\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.815471 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xxxp7\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.817190 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xxxp7\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.831470 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmsk6\" (UniqueName: \"kubernetes.io/projected/22cbc8e9-1434-4dcb-8fd3-24def21fe451-kube-api-access-jmsk6\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xxxp7\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:01:16 crc kubenswrapper[4818]: I1203 07:01:16.871592 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:01:17 crc kubenswrapper[4818]: I1203 07:01:17.399965 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7"] Dec 03 07:01:17 crc kubenswrapper[4818]: I1203 07:01:17.460607 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" event={"ID":"22cbc8e9-1434-4dcb-8fd3-24def21fe451","Type":"ContainerStarted","Data":"ad614d84ee93cd61593cd13cf89797ef9f62b3839fdffdba0e9fe24b6d0733fe"} Dec 03 07:01:18 crc kubenswrapper[4818]: I1203 07:01:18.470299 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" event={"ID":"22cbc8e9-1434-4dcb-8fd3-24def21fe451","Type":"ContainerStarted","Data":"9694c46318be95a86590bbb85b1a6555780d1d668fbff5618f9d3845259f94b3"} Dec 03 07:01:18 crc kubenswrapper[4818]: I1203 07:01:18.494861 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" podStartSLOduration=1.99539266 podStartE2EDuration="2.494841015s" podCreationTimestamp="2025-12-03 07:01:16 +0000 UTC" firstStartedPulling="2025-12-03 07:01:17.409943811 +0000 UTC m=+2035.101552563" lastFinishedPulling="2025-12-03 07:01:17.909392146 +0000 UTC m=+2035.601000918" observedRunningTime="2025-12-03 07:01:18.493296637 +0000 UTC m=+2036.184905389" watchObservedRunningTime="2025-12-03 07:01:18.494841015 +0000 UTC m=+2036.186449767" Dec 03 07:01:43 crc kubenswrapper[4818]: I1203 07:01:43.303203 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:01:43 crc kubenswrapper[4818]: I1203 07:01:43.303892 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:01:43 crc kubenswrapper[4818]: I1203 07:01:43.303956 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 07:01:43 crc kubenswrapper[4818]: I1203 07:01:43.305079 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"955318249bafbdd3f6163ed23bb008bdbe9f0304f5cc29a8cc76bbceb8f4da5b"} pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:01:43 crc kubenswrapper[4818]: I1203 07:01:43.305308 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" containerID="cri-o://955318249bafbdd3f6163ed23bb008bdbe9f0304f5cc29a8cc76bbceb8f4da5b" gracePeriod=600 Dec 03 07:01:43 crc kubenswrapper[4818]: I1203 07:01:43.897824 4818 generic.go:334] "Generic (PLEG): container finished" podID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerID="955318249bafbdd3f6163ed23bb008bdbe9f0304f5cc29a8cc76bbceb8f4da5b" exitCode=0 Dec 03 07:01:43 crc kubenswrapper[4818]: I1203 07:01:43.897886 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerDied","Data":"955318249bafbdd3f6163ed23bb008bdbe9f0304f5cc29a8cc76bbceb8f4da5b"} Dec 03 07:01:43 crc kubenswrapper[4818]: I1203 07:01:43.898197 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806"} Dec 03 07:01:43 crc kubenswrapper[4818]: I1203 07:01:43.898235 4818 scope.go:117] "RemoveContainer" containerID="cef38712982806632fe316cf0cc3134faf3e62c51472c0452be30b8c7fda5b01" Dec 03 07:02:26 crc kubenswrapper[4818]: I1203 07:02:26.321081 4818 generic.go:334] "Generic (PLEG): container finished" podID="22cbc8e9-1434-4dcb-8fd3-24def21fe451" containerID="9694c46318be95a86590bbb85b1a6555780d1d668fbff5618f9d3845259f94b3" exitCode=0 Dec 03 07:02:26 crc kubenswrapper[4818]: I1203 07:02:26.321174 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" event={"ID":"22cbc8e9-1434-4dcb-8fd3-24def21fe451","Type":"ContainerDied","Data":"9694c46318be95a86590bbb85b1a6555780d1d668fbff5618f9d3845259f94b3"} Dec 03 07:02:27 crc kubenswrapper[4818]: I1203 07:02:27.736898 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:02:27 crc kubenswrapper[4818]: I1203 07:02:27.871680 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ovncontroller-config-0\") pod \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " Dec 03 07:02:27 crc kubenswrapper[4818]: I1203 07:02:27.871721 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ssh-key\") pod \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " Dec 03 07:02:27 crc kubenswrapper[4818]: I1203 07:02:27.871754 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmsk6\" (UniqueName: \"kubernetes.io/projected/22cbc8e9-1434-4dcb-8fd3-24def21fe451-kube-api-access-jmsk6\") pod \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " Dec 03 07:02:27 crc kubenswrapper[4818]: I1203 07:02:27.871851 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-inventory\") pod \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " Dec 03 07:02:27 crc kubenswrapper[4818]: I1203 07:02:27.871923 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ovn-combined-ca-bundle\") pod \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\" (UID: \"22cbc8e9-1434-4dcb-8fd3-24def21fe451\") " Dec 03 07:02:27 crc kubenswrapper[4818]: I1203 07:02:27.878451 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "22cbc8e9-1434-4dcb-8fd3-24def21fe451" (UID: "22cbc8e9-1434-4dcb-8fd3-24def21fe451"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:02:27 crc kubenswrapper[4818]: I1203 07:02:27.879405 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22cbc8e9-1434-4dcb-8fd3-24def21fe451-kube-api-access-jmsk6" (OuterVolumeSpecName: "kube-api-access-jmsk6") pod "22cbc8e9-1434-4dcb-8fd3-24def21fe451" (UID: "22cbc8e9-1434-4dcb-8fd3-24def21fe451"). InnerVolumeSpecName "kube-api-access-jmsk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:02:27 crc kubenswrapper[4818]: I1203 07:02:27.903391 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "22cbc8e9-1434-4dcb-8fd3-24def21fe451" (UID: "22cbc8e9-1434-4dcb-8fd3-24def21fe451"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:02:27 crc kubenswrapper[4818]: I1203 07:02:27.920689 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "22cbc8e9-1434-4dcb-8fd3-24def21fe451" (UID: "22cbc8e9-1434-4dcb-8fd3-24def21fe451"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:02:27 crc kubenswrapper[4818]: I1203 07:02:27.922118 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-inventory" (OuterVolumeSpecName: "inventory") pod "22cbc8e9-1434-4dcb-8fd3-24def21fe451" (UID: "22cbc8e9-1434-4dcb-8fd3-24def21fe451"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:02:27 crc kubenswrapper[4818]: I1203 07:02:27.975409 4818 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:02:27 crc kubenswrapper[4818]: I1203 07:02:27.975453 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 07:02:27 crc kubenswrapper[4818]: I1203 07:02:27.975471 4818 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/22cbc8e9-1434-4dcb-8fd3-24def21fe451-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:02:27 crc kubenswrapper[4818]: I1203 07:02:27.975490 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmsk6\" (UniqueName: \"kubernetes.io/projected/22cbc8e9-1434-4dcb-8fd3-24def21fe451-kube-api-access-jmsk6\") on node \"crc\" DevicePath \"\"" Dec 03 07:02:27 crc kubenswrapper[4818]: I1203 07:02:27.975509 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22cbc8e9-1434-4dcb-8fd3-24def21fe451-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.359203 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" event={"ID":"22cbc8e9-1434-4dcb-8fd3-24def21fe451","Type":"ContainerDied","Data":"ad614d84ee93cd61593cd13cf89797ef9f62b3839fdffdba0e9fe24b6d0733fe"} Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.359271 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad614d84ee93cd61593cd13cf89797ef9f62b3839fdffdba0e9fe24b6d0733fe" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.359302 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xxxp7" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.462377 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9"] Dec 03 07:02:28 crc kubenswrapper[4818]: E1203 07:02:28.462804 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22cbc8e9-1434-4dcb-8fd3-24def21fe451" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.462845 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="22cbc8e9-1434-4dcb-8fd3-24def21fe451" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.463094 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="22cbc8e9-1434-4dcb-8fd3-24def21fe451" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.463740 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.466442 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.466906 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.467164 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2bpsv" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.467428 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.467626 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.468364 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.478324 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9"] Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.588804 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.588949 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqxwb\" (UniqueName: \"kubernetes.io/projected/31a69e56-fab5-4dd2-83bb-78cce024c2c1-kube-api-access-nqxwb\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.589216 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.589268 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.589459 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.589625 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.691604 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqxwb\" (UniqueName: \"kubernetes.io/projected/31a69e56-fab5-4dd2-83bb-78cce024c2c1-kube-api-access-nqxwb\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.691734 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.691778 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.691868 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.691915 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.691988 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.696127 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.696729 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.698666 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.698716 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.701200 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.710870 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqxwb\" (UniqueName: \"kubernetes.io/projected/31a69e56-fab5-4dd2-83bb-78cce024c2c1-kube-api-access-nqxwb\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:28 crc kubenswrapper[4818]: I1203 07:02:28.793272 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:02:29 crc kubenswrapper[4818]: I1203 07:02:29.357679 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9"] Dec 03 07:02:29 crc kubenswrapper[4818]: I1203 07:02:29.370773 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" event={"ID":"31a69e56-fab5-4dd2-83bb-78cce024c2c1","Type":"ContainerStarted","Data":"a4faab77b7c74a956d1118f48a0ee70a42ba2cf698e37af4b1201c7864f2b748"} Dec 03 07:02:30 crc kubenswrapper[4818]: I1203 07:02:30.384259 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" event={"ID":"31a69e56-fab5-4dd2-83bb-78cce024c2c1","Type":"ContainerStarted","Data":"b64797952f6701f0204b8278fb976ad54ff521cea2d70cf18416fe374338c814"} Dec 03 07:02:30 crc kubenswrapper[4818]: I1203 07:02:30.406333 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" podStartSLOduration=1.820312989 podStartE2EDuration="2.406295401s" podCreationTimestamp="2025-12-03 07:02:28 +0000 UTC" firstStartedPulling="2025-12-03 07:02:29.358409425 +0000 UTC m=+2107.050018187" lastFinishedPulling="2025-12-03 07:02:29.944391837 +0000 UTC m=+2107.636000599" observedRunningTime="2025-12-03 07:02:30.401937303 +0000 UTC m=+2108.093546055" watchObservedRunningTime="2025-12-03 07:02:30.406295401 +0000 UTC m=+2108.097904173" Dec 03 07:03:22 crc kubenswrapper[4818]: I1203 07:03:22.930677 4818 generic.go:334] "Generic (PLEG): container finished" podID="31a69e56-fab5-4dd2-83bb-78cce024c2c1" containerID="b64797952f6701f0204b8278fb976ad54ff521cea2d70cf18416fe374338c814" exitCode=0 Dec 03 07:03:22 crc kubenswrapper[4818]: I1203 07:03:22.930808 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" event={"ID":"31a69e56-fab5-4dd2-83bb-78cce024c2c1","Type":"ContainerDied","Data":"b64797952f6701f0204b8278fb976ad54ff521cea2d70cf18416fe374338c814"} Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.374745 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.411400 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-ssh-key\") pod \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.411540 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-neutron-ovn-metadata-agent-neutron-config-0\") pod \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.411671 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-neutron-metadata-combined-ca-bundle\") pod \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.411735 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqxwb\" (UniqueName: \"kubernetes.io/projected/31a69e56-fab5-4dd2-83bb-78cce024c2c1-kube-api-access-nqxwb\") pod \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.411771 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-nova-metadata-neutron-config-0\") pod \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.411829 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-inventory\") pod \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\" (UID: \"31a69e56-fab5-4dd2-83bb-78cce024c2c1\") " Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.426868 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "31a69e56-fab5-4dd2-83bb-78cce024c2c1" (UID: "31a69e56-fab5-4dd2-83bb-78cce024c2c1"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.441699 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31a69e56-fab5-4dd2-83bb-78cce024c2c1-kube-api-access-nqxwb" (OuterVolumeSpecName: "kube-api-access-nqxwb") pod "31a69e56-fab5-4dd2-83bb-78cce024c2c1" (UID: "31a69e56-fab5-4dd2-83bb-78cce024c2c1"). InnerVolumeSpecName "kube-api-access-nqxwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.444563 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "31a69e56-fab5-4dd2-83bb-78cce024c2c1" (UID: "31a69e56-fab5-4dd2-83bb-78cce024c2c1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.447270 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "31a69e56-fab5-4dd2-83bb-78cce024c2c1" (UID: "31a69e56-fab5-4dd2-83bb-78cce024c2c1"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.453683 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "31a69e56-fab5-4dd2-83bb-78cce024c2c1" (UID: "31a69e56-fab5-4dd2-83bb-78cce024c2c1"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.457999 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-inventory" (OuterVolumeSpecName: "inventory") pod "31a69e56-fab5-4dd2-83bb-78cce024c2c1" (UID: "31a69e56-fab5-4dd2-83bb-78cce024c2c1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.513641 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.513921 4818 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.513977 4818 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.513988 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqxwb\" (UniqueName: \"kubernetes.io/projected/31a69e56-fab5-4dd2-83bb-78cce024c2c1-kube-api-access-nqxwb\") on node \"crc\" DevicePath \"\"" Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.513997 4818 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.514010 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31a69e56-fab5-4dd2-83bb-78cce024c2c1-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.950749 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" event={"ID":"31a69e56-fab5-4dd2-83bb-78cce024c2c1","Type":"ContainerDied","Data":"a4faab77b7c74a956d1118f48a0ee70a42ba2cf698e37af4b1201c7864f2b748"} Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.950796 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4faab77b7c74a956d1118f48a0ee70a42ba2cf698e37af4b1201c7864f2b748" Dec 03 07:03:24 crc kubenswrapper[4818]: I1203 07:03:24.950877 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.071182 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd"] Dec 03 07:03:25 crc kubenswrapper[4818]: E1203 07:03:25.072222 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31a69e56-fab5-4dd2-83bb-78cce024c2c1" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.072242 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="31a69e56-fab5-4dd2-83bb-78cce024c2c1" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.072537 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="31a69e56-fab5-4dd2-83bb-78cce024c2c1" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.073367 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.076359 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.076661 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.076845 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.077030 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.077202 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2bpsv" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.083077 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd"] Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.137078 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.137167 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.137193 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7595\" (UniqueName: \"kubernetes.io/projected/e5796328-5b74-4f5a-939b-403485be16a2-kube-api-access-t7595\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.137257 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.137292 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.239510 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.240403 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.240451 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7595\" (UniqueName: \"kubernetes.io/projected/e5796328-5b74-4f5a-939b-403485be16a2-kube-api-access-t7595\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.240597 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.240653 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.244953 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.246005 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.248135 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.248397 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.258022 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7595\" (UniqueName: \"kubernetes.io/projected/e5796328-5b74-4f5a-939b-403485be16a2-kube-api-access-t7595\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:03:25 crc kubenswrapper[4818]: I1203 07:03:25.389900 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:03:26 crc kubenswrapper[4818]: I1203 07:03:26.004246 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd"] Dec 03 07:03:26 crc kubenswrapper[4818]: I1203 07:03:26.978339 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" event={"ID":"e5796328-5b74-4f5a-939b-403485be16a2","Type":"ContainerStarted","Data":"eb3936ef05c0b8dacf299ec5aab40e276a4c36802f2fc1fe76eef3624056eb68"} Dec 03 07:03:26 crc kubenswrapper[4818]: I1203 07:03:26.978703 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" event={"ID":"e5796328-5b74-4f5a-939b-403485be16a2","Type":"ContainerStarted","Data":"717ed273d52d5b2629e334549d3823db7956d5c389a1fb064ac2b19c3c53cf7f"} Dec 03 07:03:27 crc kubenswrapper[4818]: I1203 07:03:27.011918 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" podStartSLOduration=1.5249480640000002 podStartE2EDuration="2.011893179s" podCreationTimestamp="2025-12-03 07:03:25 +0000 UTC" firstStartedPulling="2025-12-03 07:03:26.002549159 +0000 UTC m=+2163.694157911" lastFinishedPulling="2025-12-03 07:03:26.489494264 +0000 UTC m=+2164.181103026" observedRunningTime="2025-12-03 07:03:27.007000487 +0000 UTC m=+2164.698609249" watchObservedRunningTime="2025-12-03 07:03:27.011893179 +0000 UTC m=+2164.703501931" Dec 03 07:03:43 crc kubenswrapper[4818]: I1203 07:03:43.301937 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:03:43 crc kubenswrapper[4818]: I1203 07:03:43.302542 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:04:05 crc kubenswrapper[4818]: I1203 07:04:05.267296 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fnq5k"] Dec 03 07:04:05 crc kubenswrapper[4818]: I1203 07:04:05.271257 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fnq5k" Dec 03 07:04:05 crc kubenswrapper[4818]: I1203 07:04:05.287392 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnq5k"] Dec 03 07:04:05 crc kubenswrapper[4818]: I1203 07:04:05.361155 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde95e0c-ea04-4142-a53c-eeac36af6783-utilities\") pod \"redhat-marketplace-fnq5k\" (UID: \"dde95e0c-ea04-4142-a53c-eeac36af6783\") " pod="openshift-marketplace/redhat-marketplace-fnq5k" Dec 03 07:04:05 crc kubenswrapper[4818]: I1203 07:04:05.361272 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde95e0c-ea04-4142-a53c-eeac36af6783-catalog-content\") pod \"redhat-marketplace-fnq5k\" (UID: \"dde95e0c-ea04-4142-a53c-eeac36af6783\") " pod="openshift-marketplace/redhat-marketplace-fnq5k" Dec 03 07:04:05 crc kubenswrapper[4818]: I1203 07:04:05.361372 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9vl8\" (UniqueName: \"kubernetes.io/projected/dde95e0c-ea04-4142-a53c-eeac36af6783-kube-api-access-j9vl8\") pod \"redhat-marketplace-fnq5k\" (UID: \"dde95e0c-ea04-4142-a53c-eeac36af6783\") " pod="openshift-marketplace/redhat-marketplace-fnq5k" Dec 03 07:04:05 crc kubenswrapper[4818]: I1203 07:04:05.463296 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde95e0c-ea04-4142-a53c-eeac36af6783-utilities\") pod \"redhat-marketplace-fnq5k\" (UID: \"dde95e0c-ea04-4142-a53c-eeac36af6783\") " pod="openshift-marketplace/redhat-marketplace-fnq5k" Dec 03 07:04:05 crc kubenswrapper[4818]: I1203 07:04:05.463414 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde95e0c-ea04-4142-a53c-eeac36af6783-catalog-content\") pod \"redhat-marketplace-fnq5k\" (UID: \"dde95e0c-ea04-4142-a53c-eeac36af6783\") " pod="openshift-marketplace/redhat-marketplace-fnq5k" Dec 03 07:04:05 crc kubenswrapper[4818]: I1203 07:04:05.463470 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9vl8\" (UniqueName: \"kubernetes.io/projected/dde95e0c-ea04-4142-a53c-eeac36af6783-kube-api-access-j9vl8\") pod \"redhat-marketplace-fnq5k\" (UID: \"dde95e0c-ea04-4142-a53c-eeac36af6783\") " pod="openshift-marketplace/redhat-marketplace-fnq5k" Dec 03 07:04:05 crc kubenswrapper[4818]: I1203 07:04:05.464408 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde95e0c-ea04-4142-a53c-eeac36af6783-utilities\") pod \"redhat-marketplace-fnq5k\" (UID: \"dde95e0c-ea04-4142-a53c-eeac36af6783\") " pod="openshift-marketplace/redhat-marketplace-fnq5k" Dec 03 07:04:05 crc kubenswrapper[4818]: I1203 07:04:05.464678 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde95e0c-ea04-4142-a53c-eeac36af6783-catalog-content\") pod \"redhat-marketplace-fnq5k\" (UID: \"dde95e0c-ea04-4142-a53c-eeac36af6783\") " pod="openshift-marketplace/redhat-marketplace-fnq5k" Dec 03 07:04:05 crc kubenswrapper[4818]: I1203 07:04:05.483910 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9vl8\" (UniqueName: \"kubernetes.io/projected/dde95e0c-ea04-4142-a53c-eeac36af6783-kube-api-access-j9vl8\") pod \"redhat-marketplace-fnq5k\" (UID: \"dde95e0c-ea04-4142-a53c-eeac36af6783\") " pod="openshift-marketplace/redhat-marketplace-fnq5k" Dec 03 07:04:05 crc kubenswrapper[4818]: I1203 07:04:05.608370 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fnq5k" Dec 03 07:04:06 crc kubenswrapper[4818]: I1203 07:04:06.125325 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnq5k"] Dec 03 07:04:06 crc kubenswrapper[4818]: I1203 07:04:06.365656 4818 generic.go:334] "Generic (PLEG): container finished" podID="dde95e0c-ea04-4142-a53c-eeac36af6783" containerID="3c85e3427d22cd2cabe8833f39c5f855632b92647f006f3a1f2aedc0f32b1b75" exitCode=0 Dec 03 07:04:06 crc kubenswrapper[4818]: I1203 07:04:06.365708 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnq5k" event={"ID":"dde95e0c-ea04-4142-a53c-eeac36af6783","Type":"ContainerDied","Data":"3c85e3427d22cd2cabe8833f39c5f855632b92647f006f3a1f2aedc0f32b1b75"} Dec 03 07:04:06 crc kubenswrapper[4818]: I1203 07:04:06.365737 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnq5k" event={"ID":"dde95e0c-ea04-4142-a53c-eeac36af6783","Type":"ContainerStarted","Data":"dd840dc52385d4e6381ccc704bb2cf46875eb2dd7a61e1d0b11445c113b6a9b1"} Dec 03 07:04:07 crc kubenswrapper[4818]: I1203 07:04:07.381531 4818 generic.go:334] "Generic (PLEG): container finished" podID="dde95e0c-ea04-4142-a53c-eeac36af6783" containerID="2d9782c3e2997e3068808946758f6462e55e59e9d34e8b52f4e6aa67ea41c7ce" exitCode=0 Dec 03 07:04:07 crc kubenswrapper[4818]: I1203 07:04:07.381651 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnq5k" event={"ID":"dde95e0c-ea04-4142-a53c-eeac36af6783","Type":"ContainerDied","Data":"2d9782c3e2997e3068808946758f6462e55e59e9d34e8b52f4e6aa67ea41c7ce"} Dec 03 07:04:08 crc kubenswrapper[4818]: I1203 07:04:08.394060 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnq5k" event={"ID":"dde95e0c-ea04-4142-a53c-eeac36af6783","Type":"ContainerStarted","Data":"5ebbd3f47c4802a88930bc30b1c37f4a1b731cdcbc18059e6681ca4ec3d5efad"} Dec 03 07:04:08 crc kubenswrapper[4818]: I1203 07:04:08.414083 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fnq5k" podStartSLOduration=1.97701693 podStartE2EDuration="3.414059717s" podCreationTimestamp="2025-12-03 07:04:05 +0000 UTC" firstStartedPulling="2025-12-03 07:04:06.367387655 +0000 UTC m=+2204.058996407" lastFinishedPulling="2025-12-03 07:04:07.804430442 +0000 UTC m=+2205.496039194" observedRunningTime="2025-12-03 07:04:08.40976715 +0000 UTC m=+2206.101375902" watchObservedRunningTime="2025-12-03 07:04:08.414059717 +0000 UTC m=+2206.105668469" Dec 03 07:04:13 crc kubenswrapper[4818]: I1203 07:04:13.302561 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:04:13 crc kubenswrapper[4818]: I1203 07:04:13.303190 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:04:15 crc kubenswrapper[4818]: I1203 07:04:15.609107 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fnq5k" Dec 03 07:04:15 crc kubenswrapper[4818]: I1203 07:04:15.609477 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fnq5k" Dec 03 07:04:15 crc kubenswrapper[4818]: I1203 07:04:15.675848 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fnq5k" Dec 03 07:04:16 crc kubenswrapper[4818]: I1203 07:04:16.523388 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fnq5k" Dec 03 07:04:16 crc kubenswrapper[4818]: I1203 07:04:16.573032 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnq5k"] Dec 03 07:04:18 crc kubenswrapper[4818]: I1203 07:04:18.500950 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fnq5k" podUID="dde95e0c-ea04-4142-a53c-eeac36af6783" containerName="registry-server" containerID="cri-o://5ebbd3f47c4802a88930bc30b1c37f4a1b731cdcbc18059e6681ca4ec3d5efad" gracePeriod=2 Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.031204 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fnq5k" Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.131156 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde95e0c-ea04-4142-a53c-eeac36af6783-utilities\") pod \"dde95e0c-ea04-4142-a53c-eeac36af6783\" (UID: \"dde95e0c-ea04-4142-a53c-eeac36af6783\") " Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.131197 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9vl8\" (UniqueName: \"kubernetes.io/projected/dde95e0c-ea04-4142-a53c-eeac36af6783-kube-api-access-j9vl8\") pod \"dde95e0c-ea04-4142-a53c-eeac36af6783\" (UID: \"dde95e0c-ea04-4142-a53c-eeac36af6783\") " Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.131522 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde95e0c-ea04-4142-a53c-eeac36af6783-catalog-content\") pod \"dde95e0c-ea04-4142-a53c-eeac36af6783\" (UID: \"dde95e0c-ea04-4142-a53c-eeac36af6783\") " Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.132110 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dde95e0c-ea04-4142-a53c-eeac36af6783-utilities" (OuterVolumeSpecName: "utilities") pod "dde95e0c-ea04-4142-a53c-eeac36af6783" (UID: "dde95e0c-ea04-4142-a53c-eeac36af6783"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.139169 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dde95e0c-ea04-4142-a53c-eeac36af6783-kube-api-access-j9vl8" (OuterVolumeSpecName: "kube-api-access-j9vl8") pod "dde95e0c-ea04-4142-a53c-eeac36af6783" (UID: "dde95e0c-ea04-4142-a53c-eeac36af6783"). InnerVolumeSpecName "kube-api-access-j9vl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.149040 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dde95e0c-ea04-4142-a53c-eeac36af6783-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dde95e0c-ea04-4142-a53c-eeac36af6783" (UID: "dde95e0c-ea04-4142-a53c-eeac36af6783"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.233876 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde95e0c-ea04-4142-a53c-eeac36af6783-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.233913 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde95e0c-ea04-4142-a53c-eeac36af6783-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.233926 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9vl8\" (UniqueName: \"kubernetes.io/projected/dde95e0c-ea04-4142-a53c-eeac36af6783-kube-api-access-j9vl8\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.509527 4818 generic.go:334] "Generic (PLEG): container finished" podID="dde95e0c-ea04-4142-a53c-eeac36af6783" containerID="5ebbd3f47c4802a88930bc30b1c37f4a1b731cdcbc18059e6681ca4ec3d5efad" exitCode=0 Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.509696 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnq5k" event={"ID":"dde95e0c-ea04-4142-a53c-eeac36af6783","Type":"ContainerDied","Data":"5ebbd3f47c4802a88930bc30b1c37f4a1b731cdcbc18059e6681ca4ec3d5efad"} Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.509796 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fnq5k" Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.509828 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fnq5k" event={"ID":"dde95e0c-ea04-4142-a53c-eeac36af6783","Type":"ContainerDied","Data":"dd840dc52385d4e6381ccc704bb2cf46875eb2dd7a61e1d0b11445c113b6a9b1"} Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.509871 4818 scope.go:117] "RemoveContainer" containerID="5ebbd3f47c4802a88930bc30b1c37f4a1b731cdcbc18059e6681ca4ec3d5efad" Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.530060 4818 scope.go:117] "RemoveContainer" containerID="2d9782c3e2997e3068808946758f6462e55e59e9d34e8b52f4e6aa67ea41c7ce" Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.550371 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnq5k"] Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.561959 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fnq5k"] Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.581736 4818 scope.go:117] "RemoveContainer" containerID="3c85e3427d22cd2cabe8833f39c5f855632b92647f006f3a1f2aedc0f32b1b75" Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.604018 4818 scope.go:117] "RemoveContainer" containerID="5ebbd3f47c4802a88930bc30b1c37f4a1b731cdcbc18059e6681ca4ec3d5efad" Dec 03 07:04:19 crc kubenswrapper[4818]: E1203 07:04:19.604489 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ebbd3f47c4802a88930bc30b1c37f4a1b731cdcbc18059e6681ca4ec3d5efad\": container with ID starting with 5ebbd3f47c4802a88930bc30b1c37f4a1b731cdcbc18059e6681ca4ec3d5efad not found: ID does not exist" containerID="5ebbd3f47c4802a88930bc30b1c37f4a1b731cdcbc18059e6681ca4ec3d5efad" Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.604522 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ebbd3f47c4802a88930bc30b1c37f4a1b731cdcbc18059e6681ca4ec3d5efad"} err="failed to get container status \"5ebbd3f47c4802a88930bc30b1c37f4a1b731cdcbc18059e6681ca4ec3d5efad\": rpc error: code = NotFound desc = could not find container \"5ebbd3f47c4802a88930bc30b1c37f4a1b731cdcbc18059e6681ca4ec3d5efad\": container with ID starting with 5ebbd3f47c4802a88930bc30b1c37f4a1b731cdcbc18059e6681ca4ec3d5efad not found: ID does not exist" Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.604542 4818 scope.go:117] "RemoveContainer" containerID="2d9782c3e2997e3068808946758f6462e55e59e9d34e8b52f4e6aa67ea41c7ce" Dec 03 07:04:19 crc kubenswrapper[4818]: E1203 07:04:19.604958 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d9782c3e2997e3068808946758f6462e55e59e9d34e8b52f4e6aa67ea41c7ce\": container with ID starting with 2d9782c3e2997e3068808946758f6462e55e59e9d34e8b52f4e6aa67ea41c7ce not found: ID does not exist" containerID="2d9782c3e2997e3068808946758f6462e55e59e9d34e8b52f4e6aa67ea41c7ce" Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.604983 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d9782c3e2997e3068808946758f6462e55e59e9d34e8b52f4e6aa67ea41c7ce"} err="failed to get container status \"2d9782c3e2997e3068808946758f6462e55e59e9d34e8b52f4e6aa67ea41c7ce\": rpc error: code = NotFound desc = could not find container \"2d9782c3e2997e3068808946758f6462e55e59e9d34e8b52f4e6aa67ea41c7ce\": container with ID starting with 2d9782c3e2997e3068808946758f6462e55e59e9d34e8b52f4e6aa67ea41c7ce not found: ID does not exist" Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.604997 4818 scope.go:117] "RemoveContainer" containerID="3c85e3427d22cd2cabe8833f39c5f855632b92647f006f3a1f2aedc0f32b1b75" Dec 03 07:04:19 crc kubenswrapper[4818]: E1203 07:04:19.605292 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c85e3427d22cd2cabe8833f39c5f855632b92647f006f3a1f2aedc0f32b1b75\": container with ID starting with 3c85e3427d22cd2cabe8833f39c5f855632b92647f006f3a1f2aedc0f32b1b75 not found: ID does not exist" containerID="3c85e3427d22cd2cabe8833f39c5f855632b92647f006f3a1f2aedc0f32b1b75" Dec 03 07:04:19 crc kubenswrapper[4818]: I1203 07:04:19.605311 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c85e3427d22cd2cabe8833f39c5f855632b92647f006f3a1f2aedc0f32b1b75"} err="failed to get container status \"3c85e3427d22cd2cabe8833f39c5f855632b92647f006f3a1f2aedc0f32b1b75\": rpc error: code = NotFound desc = could not find container \"3c85e3427d22cd2cabe8833f39c5f855632b92647f006f3a1f2aedc0f32b1b75\": container with ID starting with 3c85e3427d22cd2cabe8833f39c5f855632b92647f006f3a1f2aedc0f32b1b75 not found: ID does not exist" Dec 03 07:04:20 crc kubenswrapper[4818]: I1203 07:04:20.750980 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dde95e0c-ea04-4142-a53c-eeac36af6783" path="/var/lib/kubelet/pods/dde95e0c-ea04-4142-a53c-eeac36af6783/volumes" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.438287 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l9lwf"] Dec 03 07:04:31 crc kubenswrapper[4818]: E1203 07:04:31.439208 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dde95e0c-ea04-4142-a53c-eeac36af6783" containerName="extract-utilities" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.439226 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="dde95e0c-ea04-4142-a53c-eeac36af6783" containerName="extract-utilities" Dec 03 07:04:31 crc kubenswrapper[4818]: E1203 07:04:31.439253 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dde95e0c-ea04-4142-a53c-eeac36af6783" containerName="registry-server" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.439259 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="dde95e0c-ea04-4142-a53c-eeac36af6783" containerName="registry-server" Dec 03 07:04:31 crc kubenswrapper[4818]: E1203 07:04:31.439277 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dde95e0c-ea04-4142-a53c-eeac36af6783" containerName="extract-content" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.439282 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="dde95e0c-ea04-4142-a53c-eeac36af6783" containerName="extract-content" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.439491 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="dde95e0c-ea04-4142-a53c-eeac36af6783" containerName="registry-server" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.440946 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9lwf" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.450138 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l9lwf"] Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.527993 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77ldk\" (UniqueName: \"kubernetes.io/projected/f97cbfcf-da02-43ea-b7ef-ec690f277a19-kube-api-access-77ldk\") pod \"redhat-operators-l9lwf\" (UID: \"f97cbfcf-da02-43ea-b7ef-ec690f277a19\") " pod="openshift-marketplace/redhat-operators-l9lwf" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.528067 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f97cbfcf-da02-43ea-b7ef-ec690f277a19-utilities\") pod \"redhat-operators-l9lwf\" (UID: \"f97cbfcf-da02-43ea-b7ef-ec690f277a19\") " pod="openshift-marketplace/redhat-operators-l9lwf" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.528198 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f97cbfcf-da02-43ea-b7ef-ec690f277a19-catalog-content\") pod \"redhat-operators-l9lwf\" (UID: \"f97cbfcf-da02-43ea-b7ef-ec690f277a19\") " pod="openshift-marketplace/redhat-operators-l9lwf" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.629865 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77ldk\" (UniqueName: \"kubernetes.io/projected/f97cbfcf-da02-43ea-b7ef-ec690f277a19-kube-api-access-77ldk\") pod \"redhat-operators-l9lwf\" (UID: \"f97cbfcf-da02-43ea-b7ef-ec690f277a19\") " pod="openshift-marketplace/redhat-operators-l9lwf" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.629917 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f97cbfcf-da02-43ea-b7ef-ec690f277a19-utilities\") pod \"redhat-operators-l9lwf\" (UID: \"f97cbfcf-da02-43ea-b7ef-ec690f277a19\") " pod="openshift-marketplace/redhat-operators-l9lwf" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.629951 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f97cbfcf-da02-43ea-b7ef-ec690f277a19-catalog-content\") pod \"redhat-operators-l9lwf\" (UID: \"f97cbfcf-da02-43ea-b7ef-ec690f277a19\") " pod="openshift-marketplace/redhat-operators-l9lwf" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.630423 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f97cbfcf-da02-43ea-b7ef-ec690f277a19-utilities\") pod \"redhat-operators-l9lwf\" (UID: \"f97cbfcf-da02-43ea-b7ef-ec690f277a19\") " pod="openshift-marketplace/redhat-operators-l9lwf" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.630452 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f97cbfcf-da02-43ea-b7ef-ec690f277a19-catalog-content\") pod \"redhat-operators-l9lwf\" (UID: \"f97cbfcf-da02-43ea-b7ef-ec690f277a19\") " pod="openshift-marketplace/redhat-operators-l9lwf" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.637185 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gnsbh"] Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.639425 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gnsbh" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.646638 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gnsbh"] Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.655075 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77ldk\" (UniqueName: \"kubernetes.io/projected/f97cbfcf-da02-43ea-b7ef-ec690f277a19-kube-api-access-77ldk\") pod \"redhat-operators-l9lwf\" (UID: \"f97cbfcf-da02-43ea-b7ef-ec690f277a19\") " pod="openshift-marketplace/redhat-operators-l9lwf" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.836889 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-utilities\") pod \"community-operators-gnsbh\" (UID: \"7fd853fc-05e8-4e6b-a673-b6c6507d49f1\") " pod="openshift-marketplace/community-operators-gnsbh" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.837328 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-catalog-content\") pod \"community-operators-gnsbh\" (UID: \"7fd853fc-05e8-4e6b-a673-b6c6507d49f1\") " pod="openshift-marketplace/community-operators-gnsbh" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.837412 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbh6l\" (UniqueName: \"kubernetes.io/projected/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-kube-api-access-bbh6l\") pod \"community-operators-gnsbh\" (UID: \"7fd853fc-05e8-4e6b-a673-b6c6507d49f1\") " pod="openshift-marketplace/community-operators-gnsbh" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.841338 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9lwf" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.939323 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-catalog-content\") pod \"community-operators-gnsbh\" (UID: \"7fd853fc-05e8-4e6b-a673-b6c6507d49f1\") " pod="openshift-marketplace/community-operators-gnsbh" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.939417 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbh6l\" (UniqueName: \"kubernetes.io/projected/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-kube-api-access-bbh6l\") pod \"community-operators-gnsbh\" (UID: \"7fd853fc-05e8-4e6b-a673-b6c6507d49f1\") " pod="openshift-marketplace/community-operators-gnsbh" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.939515 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-utilities\") pod \"community-operators-gnsbh\" (UID: \"7fd853fc-05e8-4e6b-a673-b6c6507d49f1\") " pod="openshift-marketplace/community-operators-gnsbh" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.940201 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-utilities\") pod \"community-operators-gnsbh\" (UID: \"7fd853fc-05e8-4e6b-a673-b6c6507d49f1\") " pod="openshift-marketplace/community-operators-gnsbh" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.940507 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-catalog-content\") pod \"community-operators-gnsbh\" (UID: \"7fd853fc-05e8-4e6b-a673-b6c6507d49f1\") " pod="openshift-marketplace/community-operators-gnsbh" Dec 03 07:04:31 crc kubenswrapper[4818]: I1203 07:04:31.965749 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbh6l\" (UniqueName: \"kubernetes.io/projected/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-kube-api-access-bbh6l\") pod \"community-operators-gnsbh\" (UID: \"7fd853fc-05e8-4e6b-a673-b6c6507d49f1\") " pod="openshift-marketplace/community-operators-gnsbh" Dec 03 07:04:32 crc kubenswrapper[4818]: I1203 07:04:32.004439 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gnsbh" Dec 03 07:04:32 crc kubenswrapper[4818]: I1203 07:04:32.436968 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l9lwf"] Dec 03 07:04:32 crc kubenswrapper[4818]: I1203 07:04:32.616618 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gnsbh"] Dec 03 07:04:32 crc kubenswrapper[4818]: W1203 07:04:32.625358 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fd853fc_05e8_4e6b_a673_b6c6507d49f1.slice/crio-a5bbd25a4a5920d560449e8a3b205fc3395cd0cd44af4abe6e77711540b7b377 WatchSource:0}: Error finding container a5bbd25a4a5920d560449e8a3b205fc3395cd0cd44af4abe6e77711540b7b377: Status 404 returned error can't find the container with id a5bbd25a4a5920d560449e8a3b205fc3395cd0cd44af4abe6e77711540b7b377 Dec 03 07:04:32 crc kubenswrapper[4818]: I1203 07:04:32.639265 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9lwf" event={"ID":"f97cbfcf-da02-43ea-b7ef-ec690f277a19","Type":"ContainerStarted","Data":"51636e7894a12721048ce0e3cf9747fc26018c830ee1db6a9dc7229d5e1ee0cc"} Dec 03 07:04:33 crc kubenswrapper[4818]: I1203 07:04:33.650696 4818 generic.go:334] "Generic (PLEG): container finished" podID="f97cbfcf-da02-43ea-b7ef-ec690f277a19" containerID="d66d1812d6072d3b8aac006909a4c3bd42b8d933bbd1bc76270c9246d66d090a" exitCode=0 Dec 03 07:04:33 crc kubenswrapper[4818]: I1203 07:04:33.650848 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9lwf" event={"ID":"f97cbfcf-da02-43ea-b7ef-ec690f277a19","Type":"ContainerDied","Data":"d66d1812d6072d3b8aac006909a4c3bd42b8d933bbd1bc76270c9246d66d090a"} Dec 03 07:04:33 crc kubenswrapper[4818]: I1203 07:04:33.656124 4818 generic.go:334] "Generic (PLEG): container finished" podID="7fd853fc-05e8-4e6b-a673-b6c6507d49f1" containerID="3c543ab59de453fc96850ae0e7d1a62af51682acb1260e32fb21d098a47c5d8d" exitCode=0 Dec 03 07:04:33 crc kubenswrapper[4818]: I1203 07:04:33.656275 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnsbh" event={"ID":"7fd853fc-05e8-4e6b-a673-b6c6507d49f1","Type":"ContainerDied","Data":"3c543ab59de453fc96850ae0e7d1a62af51682acb1260e32fb21d098a47c5d8d"} Dec 03 07:04:33 crc kubenswrapper[4818]: I1203 07:04:33.656399 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnsbh" event={"ID":"7fd853fc-05e8-4e6b-a673-b6c6507d49f1","Type":"ContainerStarted","Data":"a5bbd25a4a5920d560449e8a3b205fc3395cd0cd44af4abe6e77711540b7b377"} Dec 03 07:04:35 crc kubenswrapper[4818]: I1203 07:04:35.682330 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9lwf" event={"ID":"f97cbfcf-da02-43ea-b7ef-ec690f277a19","Type":"ContainerStarted","Data":"7f513f1db2fd852a2cde2cf76db930c1f43701fc027311fa41f90aeb6de63e38"} Dec 03 07:04:35 crc kubenswrapper[4818]: I1203 07:04:35.687448 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnsbh" event={"ID":"7fd853fc-05e8-4e6b-a673-b6c6507d49f1","Type":"ContainerStarted","Data":"9d593a1aca0ceeefeca5d25b06f24f25e712627a80b4183e4ed87d905b122e5d"} Dec 03 07:04:38 crc kubenswrapper[4818]: I1203 07:04:38.723326 4818 generic.go:334] "Generic (PLEG): container finished" podID="f97cbfcf-da02-43ea-b7ef-ec690f277a19" containerID="7f513f1db2fd852a2cde2cf76db930c1f43701fc027311fa41f90aeb6de63e38" exitCode=0 Dec 03 07:04:38 crc kubenswrapper[4818]: I1203 07:04:38.723383 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9lwf" event={"ID":"f97cbfcf-da02-43ea-b7ef-ec690f277a19","Type":"ContainerDied","Data":"7f513f1db2fd852a2cde2cf76db930c1f43701fc027311fa41f90aeb6de63e38"} Dec 03 07:04:38 crc kubenswrapper[4818]: I1203 07:04:38.732063 4818 generic.go:334] "Generic (PLEG): container finished" podID="7fd853fc-05e8-4e6b-a673-b6c6507d49f1" containerID="9d593a1aca0ceeefeca5d25b06f24f25e712627a80b4183e4ed87d905b122e5d" exitCode=0 Dec 03 07:04:38 crc kubenswrapper[4818]: I1203 07:04:38.732189 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnsbh" event={"ID":"7fd853fc-05e8-4e6b-a673-b6c6507d49f1","Type":"ContainerDied","Data":"9d593a1aca0ceeefeca5d25b06f24f25e712627a80b4183e4ed87d905b122e5d"} Dec 03 07:04:40 crc kubenswrapper[4818]: I1203 07:04:40.758981 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnsbh" event={"ID":"7fd853fc-05e8-4e6b-a673-b6c6507d49f1","Type":"ContainerStarted","Data":"242b1abc33d756593ff98c71eb9f9566a6cc4bb01a3cb30fc951ecaf32aca700"} Dec 03 07:04:40 crc kubenswrapper[4818]: I1203 07:04:40.759749 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9lwf" event={"ID":"f97cbfcf-da02-43ea-b7ef-ec690f277a19","Type":"ContainerStarted","Data":"aa2d512f3e5b979b0ab73c1ddc298500be4da21ef5253cd3833c40a2c33d5195"} Dec 03 07:04:40 crc kubenswrapper[4818]: I1203 07:04:40.792662 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l9lwf" podStartSLOduration=3.650589475 podStartE2EDuration="9.790552725s" podCreationTimestamp="2025-12-03 07:04:31 +0000 UTC" firstStartedPulling="2025-12-03 07:04:33.65292538 +0000 UTC m=+2231.344534132" lastFinishedPulling="2025-12-03 07:04:39.79288861 +0000 UTC m=+2237.484497382" observedRunningTime="2025-12-03 07:04:40.773805469 +0000 UTC m=+2238.465414221" watchObservedRunningTime="2025-12-03 07:04:40.790552725 +0000 UTC m=+2238.482161507" Dec 03 07:04:40 crc kubenswrapper[4818]: I1203 07:04:40.803380 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gnsbh" podStartSLOduration=3.78000601 podStartE2EDuration="9.803355753s" podCreationTimestamp="2025-12-03 07:04:31 +0000 UTC" firstStartedPulling="2025-12-03 07:04:33.659653157 +0000 UTC m=+2231.351261919" lastFinishedPulling="2025-12-03 07:04:39.68300288 +0000 UTC m=+2237.374611662" observedRunningTime="2025-12-03 07:04:40.795009735 +0000 UTC m=+2238.486618487" watchObservedRunningTime="2025-12-03 07:04:40.803355753 +0000 UTC m=+2238.494964505" Dec 03 07:04:41 crc kubenswrapper[4818]: I1203 07:04:41.843544 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l9lwf" Dec 03 07:04:41 crc kubenswrapper[4818]: I1203 07:04:41.843587 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l9lwf" Dec 03 07:04:42 crc kubenswrapper[4818]: I1203 07:04:42.005175 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gnsbh" Dec 03 07:04:42 crc kubenswrapper[4818]: I1203 07:04:42.005214 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gnsbh" Dec 03 07:04:42 crc kubenswrapper[4818]: I1203 07:04:42.051194 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gnsbh" Dec 03 07:04:42 crc kubenswrapper[4818]: I1203 07:04:42.887107 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l9lwf" podUID="f97cbfcf-da02-43ea-b7ef-ec690f277a19" containerName="registry-server" probeResult="failure" output=< Dec 03 07:04:42 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Dec 03 07:04:42 crc kubenswrapper[4818]: > Dec 03 07:04:43 crc kubenswrapper[4818]: I1203 07:04:43.302846 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:04:43 crc kubenswrapper[4818]: I1203 07:04:43.302946 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:04:43 crc kubenswrapper[4818]: I1203 07:04:43.303009 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 07:04:43 crc kubenswrapper[4818]: I1203 07:04:43.304127 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806"} pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:04:43 crc kubenswrapper[4818]: I1203 07:04:43.304225 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" containerID="cri-o://46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" gracePeriod=600 Dec 03 07:04:43 crc kubenswrapper[4818]: I1203 07:04:43.784329 4818 generic.go:334] "Generic (PLEG): container finished" podID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" exitCode=0 Dec 03 07:04:43 crc kubenswrapper[4818]: I1203 07:04:43.784382 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerDied","Data":"46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806"} Dec 03 07:04:43 crc kubenswrapper[4818]: I1203 07:04:43.784420 4818 scope.go:117] "RemoveContainer" containerID="955318249bafbdd3f6163ed23bb008bdbe9f0304f5cc29a8cc76bbceb8f4da5b" Dec 03 07:04:43 crc kubenswrapper[4818]: E1203 07:04:43.934008 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:04:44 crc kubenswrapper[4818]: I1203 07:04:44.796908 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:04:44 crc kubenswrapper[4818]: E1203 07:04:44.797507 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:04:51 crc kubenswrapper[4818]: I1203 07:04:51.915660 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l9lwf" Dec 03 07:04:51 crc kubenswrapper[4818]: I1203 07:04:51.987468 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l9lwf" Dec 03 07:04:52 crc kubenswrapper[4818]: I1203 07:04:52.056380 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gnsbh" Dec 03 07:04:52 crc kubenswrapper[4818]: I1203 07:04:52.169374 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l9lwf"] Dec 03 07:04:53 crc kubenswrapper[4818]: I1203 07:04:53.888157 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l9lwf" podUID="f97cbfcf-da02-43ea-b7ef-ec690f277a19" containerName="registry-server" containerID="cri-o://aa2d512f3e5b979b0ab73c1ddc298500be4da21ef5253cd3833c40a2c33d5195" gracePeriod=2 Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.368773 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gnsbh"] Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.369331 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gnsbh" podUID="7fd853fc-05e8-4e6b-a673-b6c6507d49f1" containerName="registry-server" containerID="cri-o://242b1abc33d756593ff98c71eb9f9566a6cc4bb01a3cb30fc951ecaf32aca700" gracePeriod=2 Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.530399 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9lwf" Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.701611 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f97cbfcf-da02-43ea-b7ef-ec690f277a19-utilities\") pod \"f97cbfcf-da02-43ea-b7ef-ec690f277a19\" (UID: \"f97cbfcf-da02-43ea-b7ef-ec690f277a19\") " Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.701742 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f97cbfcf-da02-43ea-b7ef-ec690f277a19-catalog-content\") pod \"f97cbfcf-da02-43ea-b7ef-ec690f277a19\" (UID: \"f97cbfcf-da02-43ea-b7ef-ec690f277a19\") " Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.701930 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77ldk\" (UniqueName: \"kubernetes.io/projected/f97cbfcf-da02-43ea-b7ef-ec690f277a19-kube-api-access-77ldk\") pod \"f97cbfcf-da02-43ea-b7ef-ec690f277a19\" (UID: \"f97cbfcf-da02-43ea-b7ef-ec690f277a19\") " Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.702631 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f97cbfcf-da02-43ea-b7ef-ec690f277a19-utilities" (OuterVolumeSpecName: "utilities") pod "f97cbfcf-da02-43ea-b7ef-ec690f277a19" (UID: "f97cbfcf-da02-43ea-b7ef-ec690f277a19"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.714670 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f97cbfcf-da02-43ea-b7ef-ec690f277a19-kube-api-access-77ldk" (OuterVolumeSpecName: "kube-api-access-77ldk") pod "f97cbfcf-da02-43ea-b7ef-ec690f277a19" (UID: "f97cbfcf-da02-43ea-b7ef-ec690f277a19"). InnerVolumeSpecName "kube-api-access-77ldk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.772070 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gnsbh" Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.804346 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77ldk\" (UniqueName: \"kubernetes.io/projected/f97cbfcf-da02-43ea-b7ef-ec690f277a19-kube-api-access-77ldk\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.804638 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f97cbfcf-da02-43ea-b7ef-ec690f277a19-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.824699 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f97cbfcf-da02-43ea-b7ef-ec690f277a19-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f97cbfcf-da02-43ea-b7ef-ec690f277a19" (UID: "f97cbfcf-da02-43ea-b7ef-ec690f277a19"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.903744 4818 generic.go:334] "Generic (PLEG): container finished" podID="f97cbfcf-da02-43ea-b7ef-ec690f277a19" containerID="aa2d512f3e5b979b0ab73c1ddc298500be4da21ef5253cd3833c40a2c33d5195" exitCode=0 Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.903802 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9lwf" event={"ID":"f97cbfcf-da02-43ea-b7ef-ec690f277a19","Type":"ContainerDied","Data":"aa2d512f3e5b979b0ab73c1ddc298500be4da21ef5253cd3833c40a2c33d5195"} Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.903871 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9lwf" event={"ID":"f97cbfcf-da02-43ea-b7ef-ec690f277a19","Type":"ContainerDied","Data":"51636e7894a12721048ce0e3cf9747fc26018c830ee1db6a9dc7229d5e1ee0cc"} Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.903880 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9lwf" Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.903907 4818 scope.go:117] "RemoveContainer" containerID="aa2d512f3e5b979b0ab73c1ddc298500be4da21ef5253cd3833c40a2c33d5195" Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.905786 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbh6l\" (UniqueName: \"kubernetes.io/projected/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-kube-api-access-bbh6l\") pod \"7fd853fc-05e8-4e6b-a673-b6c6507d49f1\" (UID: \"7fd853fc-05e8-4e6b-a673-b6c6507d49f1\") " Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.906285 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-utilities\") pod \"7fd853fc-05e8-4e6b-a673-b6c6507d49f1\" (UID: \"7fd853fc-05e8-4e6b-a673-b6c6507d49f1\") " Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.906345 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-catalog-content\") pod \"7fd853fc-05e8-4e6b-a673-b6c6507d49f1\" (UID: \"7fd853fc-05e8-4e6b-a673-b6c6507d49f1\") " Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.906899 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f97cbfcf-da02-43ea-b7ef-ec690f277a19-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.911106 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-utilities" (OuterVolumeSpecName: "utilities") pod "7fd853fc-05e8-4e6b-a673-b6c6507d49f1" (UID: "7fd853fc-05e8-4e6b-a673-b6c6507d49f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.911150 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-kube-api-access-bbh6l" (OuterVolumeSpecName: "kube-api-access-bbh6l") pod "7fd853fc-05e8-4e6b-a673-b6c6507d49f1" (UID: "7fd853fc-05e8-4e6b-a673-b6c6507d49f1"). InnerVolumeSpecName "kube-api-access-bbh6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.913391 4818 generic.go:334] "Generic (PLEG): container finished" podID="7fd853fc-05e8-4e6b-a673-b6c6507d49f1" containerID="242b1abc33d756593ff98c71eb9f9566a6cc4bb01a3cb30fc951ecaf32aca700" exitCode=0 Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.913661 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnsbh" event={"ID":"7fd853fc-05e8-4e6b-a673-b6c6507d49f1","Type":"ContainerDied","Data":"242b1abc33d756593ff98c71eb9f9566a6cc4bb01a3cb30fc951ecaf32aca700"} Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.913701 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnsbh" event={"ID":"7fd853fc-05e8-4e6b-a673-b6c6507d49f1","Type":"ContainerDied","Data":"a5bbd25a4a5920d560449e8a3b205fc3395cd0cd44af4abe6e77711540b7b377"} Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.913959 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gnsbh" Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.934025 4818 scope.go:117] "RemoveContainer" containerID="7f513f1db2fd852a2cde2cf76db930c1f43701fc027311fa41f90aeb6de63e38" Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.965969 4818 scope.go:117] "RemoveContainer" containerID="d66d1812d6072d3b8aac006909a4c3bd42b8d933bbd1bc76270c9246d66d090a" Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.974313 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l9lwf"] Dec 03 07:04:54 crc kubenswrapper[4818]: I1203 07:04:54.983471 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l9lwf"] Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.000805 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fd853fc-05e8-4e6b-a673-b6c6507d49f1" (UID: "7fd853fc-05e8-4e6b-a673-b6c6507d49f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.000875 4818 scope.go:117] "RemoveContainer" containerID="aa2d512f3e5b979b0ab73c1ddc298500be4da21ef5253cd3833c40a2c33d5195" Dec 03 07:04:55 crc kubenswrapper[4818]: E1203 07:04:55.001400 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa2d512f3e5b979b0ab73c1ddc298500be4da21ef5253cd3833c40a2c33d5195\": container with ID starting with aa2d512f3e5b979b0ab73c1ddc298500be4da21ef5253cd3833c40a2c33d5195 not found: ID does not exist" containerID="aa2d512f3e5b979b0ab73c1ddc298500be4da21ef5253cd3833c40a2c33d5195" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.001443 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa2d512f3e5b979b0ab73c1ddc298500be4da21ef5253cd3833c40a2c33d5195"} err="failed to get container status \"aa2d512f3e5b979b0ab73c1ddc298500be4da21ef5253cd3833c40a2c33d5195\": rpc error: code = NotFound desc = could not find container \"aa2d512f3e5b979b0ab73c1ddc298500be4da21ef5253cd3833c40a2c33d5195\": container with ID starting with aa2d512f3e5b979b0ab73c1ddc298500be4da21ef5253cd3833c40a2c33d5195 not found: ID does not exist" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.001469 4818 scope.go:117] "RemoveContainer" containerID="7f513f1db2fd852a2cde2cf76db930c1f43701fc027311fa41f90aeb6de63e38" Dec 03 07:04:55 crc kubenswrapper[4818]: E1203 07:04:55.001973 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f513f1db2fd852a2cde2cf76db930c1f43701fc027311fa41f90aeb6de63e38\": container with ID starting with 7f513f1db2fd852a2cde2cf76db930c1f43701fc027311fa41f90aeb6de63e38 not found: ID does not exist" containerID="7f513f1db2fd852a2cde2cf76db930c1f43701fc027311fa41f90aeb6de63e38" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.001995 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f513f1db2fd852a2cde2cf76db930c1f43701fc027311fa41f90aeb6de63e38"} err="failed to get container status \"7f513f1db2fd852a2cde2cf76db930c1f43701fc027311fa41f90aeb6de63e38\": rpc error: code = NotFound desc = could not find container \"7f513f1db2fd852a2cde2cf76db930c1f43701fc027311fa41f90aeb6de63e38\": container with ID starting with 7f513f1db2fd852a2cde2cf76db930c1f43701fc027311fa41f90aeb6de63e38 not found: ID does not exist" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.002010 4818 scope.go:117] "RemoveContainer" containerID="d66d1812d6072d3b8aac006909a4c3bd42b8d933bbd1bc76270c9246d66d090a" Dec 03 07:04:55 crc kubenswrapper[4818]: E1203 07:04:55.002324 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d66d1812d6072d3b8aac006909a4c3bd42b8d933bbd1bc76270c9246d66d090a\": container with ID starting with d66d1812d6072d3b8aac006909a4c3bd42b8d933bbd1bc76270c9246d66d090a not found: ID does not exist" containerID="d66d1812d6072d3b8aac006909a4c3bd42b8d933bbd1bc76270c9246d66d090a" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.002361 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d66d1812d6072d3b8aac006909a4c3bd42b8d933bbd1bc76270c9246d66d090a"} err="failed to get container status \"d66d1812d6072d3b8aac006909a4c3bd42b8d933bbd1bc76270c9246d66d090a\": rpc error: code = NotFound desc = could not find container \"d66d1812d6072d3b8aac006909a4c3bd42b8d933bbd1bc76270c9246d66d090a\": container with ID starting with d66d1812d6072d3b8aac006909a4c3bd42b8d933bbd1bc76270c9246d66d090a not found: ID does not exist" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.002389 4818 scope.go:117] "RemoveContainer" containerID="242b1abc33d756593ff98c71eb9f9566a6cc4bb01a3cb30fc951ecaf32aca700" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.010131 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.010167 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.010183 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbh6l\" (UniqueName: \"kubernetes.io/projected/7fd853fc-05e8-4e6b-a673-b6c6507d49f1-kube-api-access-bbh6l\") on node \"crc\" DevicePath \"\"" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.020388 4818 scope.go:117] "RemoveContainer" containerID="9d593a1aca0ceeefeca5d25b06f24f25e712627a80b4183e4ed87d905b122e5d" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.042743 4818 scope.go:117] "RemoveContainer" containerID="3c543ab59de453fc96850ae0e7d1a62af51682acb1260e32fb21d098a47c5d8d" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.060311 4818 scope.go:117] "RemoveContainer" containerID="242b1abc33d756593ff98c71eb9f9566a6cc4bb01a3cb30fc951ecaf32aca700" Dec 03 07:04:55 crc kubenswrapper[4818]: E1203 07:04:55.061182 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"242b1abc33d756593ff98c71eb9f9566a6cc4bb01a3cb30fc951ecaf32aca700\": container with ID starting with 242b1abc33d756593ff98c71eb9f9566a6cc4bb01a3cb30fc951ecaf32aca700 not found: ID does not exist" containerID="242b1abc33d756593ff98c71eb9f9566a6cc4bb01a3cb30fc951ecaf32aca700" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.061238 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"242b1abc33d756593ff98c71eb9f9566a6cc4bb01a3cb30fc951ecaf32aca700"} err="failed to get container status \"242b1abc33d756593ff98c71eb9f9566a6cc4bb01a3cb30fc951ecaf32aca700\": rpc error: code = NotFound desc = could not find container \"242b1abc33d756593ff98c71eb9f9566a6cc4bb01a3cb30fc951ecaf32aca700\": container with ID starting with 242b1abc33d756593ff98c71eb9f9566a6cc4bb01a3cb30fc951ecaf32aca700 not found: ID does not exist" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.061272 4818 scope.go:117] "RemoveContainer" containerID="9d593a1aca0ceeefeca5d25b06f24f25e712627a80b4183e4ed87d905b122e5d" Dec 03 07:04:55 crc kubenswrapper[4818]: E1203 07:04:55.061569 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d593a1aca0ceeefeca5d25b06f24f25e712627a80b4183e4ed87d905b122e5d\": container with ID starting with 9d593a1aca0ceeefeca5d25b06f24f25e712627a80b4183e4ed87d905b122e5d not found: ID does not exist" containerID="9d593a1aca0ceeefeca5d25b06f24f25e712627a80b4183e4ed87d905b122e5d" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.061616 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d593a1aca0ceeefeca5d25b06f24f25e712627a80b4183e4ed87d905b122e5d"} err="failed to get container status \"9d593a1aca0ceeefeca5d25b06f24f25e712627a80b4183e4ed87d905b122e5d\": rpc error: code = NotFound desc = could not find container \"9d593a1aca0ceeefeca5d25b06f24f25e712627a80b4183e4ed87d905b122e5d\": container with ID starting with 9d593a1aca0ceeefeca5d25b06f24f25e712627a80b4183e4ed87d905b122e5d not found: ID does not exist" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.061642 4818 scope.go:117] "RemoveContainer" containerID="3c543ab59de453fc96850ae0e7d1a62af51682acb1260e32fb21d098a47c5d8d" Dec 03 07:04:55 crc kubenswrapper[4818]: E1203 07:04:55.061862 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c543ab59de453fc96850ae0e7d1a62af51682acb1260e32fb21d098a47c5d8d\": container with ID starting with 3c543ab59de453fc96850ae0e7d1a62af51682acb1260e32fb21d098a47c5d8d not found: ID does not exist" containerID="3c543ab59de453fc96850ae0e7d1a62af51682acb1260e32fb21d098a47c5d8d" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.061892 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c543ab59de453fc96850ae0e7d1a62af51682acb1260e32fb21d098a47c5d8d"} err="failed to get container status \"3c543ab59de453fc96850ae0e7d1a62af51682acb1260e32fb21d098a47c5d8d\": rpc error: code = NotFound desc = could not find container \"3c543ab59de453fc96850ae0e7d1a62af51682acb1260e32fb21d098a47c5d8d\": container with ID starting with 3c543ab59de453fc96850ae0e7d1a62af51682acb1260e32fb21d098a47c5d8d not found: ID does not exist" Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.249105 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gnsbh"] Dec 03 07:04:55 crc kubenswrapper[4818]: I1203 07:04:55.258184 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gnsbh"] Dec 03 07:04:56 crc kubenswrapper[4818]: I1203 07:04:56.751051 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fd853fc-05e8-4e6b-a673-b6c6507d49f1" path="/var/lib/kubelet/pods/7fd853fc-05e8-4e6b-a673-b6c6507d49f1/volumes" Dec 03 07:04:56 crc kubenswrapper[4818]: I1203 07:04:56.752445 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f97cbfcf-da02-43ea-b7ef-ec690f277a19" path="/var/lib/kubelet/pods/f97cbfcf-da02-43ea-b7ef-ec690f277a19/volumes" Dec 03 07:04:57 crc kubenswrapper[4818]: I1203 07:04:57.738871 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:04:57 crc kubenswrapper[4818]: E1203 07:04:57.739477 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:05:07 crc kubenswrapper[4818]: I1203 07:05:07.797107 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wfr2r"] Dec 03 07:05:07 crc kubenswrapper[4818]: E1203 07:05:07.798291 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f97cbfcf-da02-43ea-b7ef-ec690f277a19" containerName="registry-server" Dec 03 07:05:07 crc kubenswrapper[4818]: I1203 07:05:07.798314 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f97cbfcf-da02-43ea-b7ef-ec690f277a19" containerName="registry-server" Dec 03 07:05:07 crc kubenswrapper[4818]: E1203 07:05:07.798340 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f97cbfcf-da02-43ea-b7ef-ec690f277a19" containerName="extract-utilities" Dec 03 07:05:07 crc kubenswrapper[4818]: I1203 07:05:07.798350 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f97cbfcf-da02-43ea-b7ef-ec690f277a19" containerName="extract-utilities" Dec 03 07:05:07 crc kubenswrapper[4818]: E1203 07:05:07.798366 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f97cbfcf-da02-43ea-b7ef-ec690f277a19" containerName="extract-content" Dec 03 07:05:07 crc kubenswrapper[4818]: I1203 07:05:07.798377 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="f97cbfcf-da02-43ea-b7ef-ec690f277a19" containerName="extract-content" Dec 03 07:05:07 crc kubenswrapper[4818]: E1203 07:05:07.798400 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd853fc-05e8-4e6b-a673-b6c6507d49f1" containerName="extract-content" Dec 03 07:05:07 crc kubenswrapper[4818]: I1203 07:05:07.798410 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd853fc-05e8-4e6b-a673-b6c6507d49f1" containerName="extract-content" Dec 03 07:05:07 crc kubenswrapper[4818]: E1203 07:05:07.798435 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd853fc-05e8-4e6b-a673-b6c6507d49f1" containerName="registry-server" Dec 03 07:05:07 crc kubenswrapper[4818]: I1203 07:05:07.798444 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd853fc-05e8-4e6b-a673-b6c6507d49f1" containerName="registry-server" Dec 03 07:05:07 crc kubenswrapper[4818]: E1203 07:05:07.798462 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd853fc-05e8-4e6b-a673-b6c6507d49f1" containerName="extract-utilities" Dec 03 07:05:07 crc kubenswrapper[4818]: I1203 07:05:07.798471 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd853fc-05e8-4e6b-a673-b6c6507d49f1" containerName="extract-utilities" Dec 03 07:05:07 crc kubenswrapper[4818]: I1203 07:05:07.798902 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fd853fc-05e8-4e6b-a673-b6c6507d49f1" containerName="registry-server" Dec 03 07:05:07 crc kubenswrapper[4818]: I1203 07:05:07.798924 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="f97cbfcf-da02-43ea-b7ef-ec690f277a19" containerName="registry-server" Dec 03 07:05:07 crc kubenswrapper[4818]: I1203 07:05:07.801542 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfr2r" Dec 03 07:05:07 crc kubenswrapper[4818]: I1203 07:05:07.814924 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wfr2r"] Dec 03 07:05:07 crc kubenswrapper[4818]: I1203 07:05:07.935198 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fdae954-7222-4ebd-8b3a-205a9fc21094-utilities\") pod \"certified-operators-wfr2r\" (UID: \"4fdae954-7222-4ebd-8b3a-205a9fc21094\") " pod="openshift-marketplace/certified-operators-wfr2r" Dec 03 07:05:07 crc kubenswrapper[4818]: I1203 07:05:07.935280 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fdae954-7222-4ebd-8b3a-205a9fc21094-catalog-content\") pod \"certified-operators-wfr2r\" (UID: \"4fdae954-7222-4ebd-8b3a-205a9fc21094\") " pod="openshift-marketplace/certified-operators-wfr2r" Dec 03 07:05:07 crc kubenswrapper[4818]: I1203 07:05:07.935379 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbn9m\" (UniqueName: \"kubernetes.io/projected/4fdae954-7222-4ebd-8b3a-205a9fc21094-kube-api-access-cbn9m\") pod \"certified-operators-wfr2r\" (UID: \"4fdae954-7222-4ebd-8b3a-205a9fc21094\") " pod="openshift-marketplace/certified-operators-wfr2r" Dec 03 07:05:08 crc kubenswrapper[4818]: I1203 07:05:08.037009 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fdae954-7222-4ebd-8b3a-205a9fc21094-utilities\") pod \"certified-operators-wfr2r\" (UID: \"4fdae954-7222-4ebd-8b3a-205a9fc21094\") " pod="openshift-marketplace/certified-operators-wfr2r" Dec 03 07:05:08 crc kubenswrapper[4818]: I1203 07:05:08.037282 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fdae954-7222-4ebd-8b3a-205a9fc21094-catalog-content\") pod \"certified-operators-wfr2r\" (UID: \"4fdae954-7222-4ebd-8b3a-205a9fc21094\") " pod="openshift-marketplace/certified-operators-wfr2r" Dec 03 07:05:08 crc kubenswrapper[4818]: I1203 07:05:08.037404 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbn9m\" (UniqueName: \"kubernetes.io/projected/4fdae954-7222-4ebd-8b3a-205a9fc21094-kube-api-access-cbn9m\") pod \"certified-operators-wfr2r\" (UID: \"4fdae954-7222-4ebd-8b3a-205a9fc21094\") " pod="openshift-marketplace/certified-operators-wfr2r" Dec 03 07:05:08 crc kubenswrapper[4818]: I1203 07:05:08.037679 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fdae954-7222-4ebd-8b3a-205a9fc21094-utilities\") pod \"certified-operators-wfr2r\" (UID: \"4fdae954-7222-4ebd-8b3a-205a9fc21094\") " pod="openshift-marketplace/certified-operators-wfr2r" Dec 03 07:05:08 crc kubenswrapper[4818]: I1203 07:05:08.037953 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fdae954-7222-4ebd-8b3a-205a9fc21094-catalog-content\") pod \"certified-operators-wfr2r\" (UID: \"4fdae954-7222-4ebd-8b3a-205a9fc21094\") " pod="openshift-marketplace/certified-operators-wfr2r" Dec 03 07:05:08 crc kubenswrapper[4818]: I1203 07:05:08.061295 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbn9m\" (UniqueName: \"kubernetes.io/projected/4fdae954-7222-4ebd-8b3a-205a9fc21094-kube-api-access-cbn9m\") pod \"certified-operators-wfr2r\" (UID: \"4fdae954-7222-4ebd-8b3a-205a9fc21094\") " pod="openshift-marketplace/certified-operators-wfr2r" Dec 03 07:05:08 crc kubenswrapper[4818]: I1203 07:05:08.124245 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfr2r" Dec 03 07:05:08 crc kubenswrapper[4818]: I1203 07:05:08.622935 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wfr2r"] Dec 03 07:05:09 crc kubenswrapper[4818]: I1203 07:05:09.059314 4818 generic.go:334] "Generic (PLEG): container finished" podID="4fdae954-7222-4ebd-8b3a-205a9fc21094" containerID="c9373927d9ab3148c6c8f0926d5f28a82e0d7e4663f087c6cfd2ebb64b9bfe7d" exitCode=0 Dec 03 07:05:09 crc kubenswrapper[4818]: I1203 07:05:09.059370 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfr2r" event={"ID":"4fdae954-7222-4ebd-8b3a-205a9fc21094","Type":"ContainerDied","Data":"c9373927d9ab3148c6c8f0926d5f28a82e0d7e4663f087c6cfd2ebb64b9bfe7d"} Dec 03 07:05:09 crc kubenswrapper[4818]: I1203 07:05:09.059638 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfr2r" event={"ID":"4fdae954-7222-4ebd-8b3a-205a9fc21094","Type":"ContainerStarted","Data":"864bd65f590e2286349a6e1d7871795e2c7379dc374feb49d7cf28e57759776a"} Dec 03 07:05:09 crc kubenswrapper[4818]: I1203 07:05:09.061192 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 07:05:09 crc kubenswrapper[4818]: I1203 07:05:09.743654 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:05:09 crc kubenswrapper[4818]: E1203 07:05:09.744715 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:05:10 crc kubenswrapper[4818]: I1203 07:05:10.077463 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfr2r" event={"ID":"4fdae954-7222-4ebd-8b3a-205a9fc21094","Type":"ContainerStarted","Data":"53f33a94a232e411ce813b77503be6fed91f737eb9b262a036ceb71eda3d99d1"} Dec 03 07:05:11 crc kubenswrapper[4818]: I1203 07:05:11.133695 4818 generic.go:334] "Generic (PLEG): container finished" podID="4fdae954-7222-4ebd-8b3a-205a9fc21094" containerID="53f33a94a232e411ce813b77503be6fed91f737eb9b262a036ceb71eda3d99d1" exitCode=0 Dec 03 07:05:11 crc kubenswrapper[4818]: I1203 07:05:11.133781 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfr2r" event={"ID":"4fdae954-7222-4ebd-8b3a-205a9fc21094","Type":"ContainerDied","Data":"53f33a94a232e411ce813b77503be6fed91f737eb9b262a036ceb71eda3d99d1"} Dec 03 07:05:12 crc kubenswrapper[4818]: I1203 07:05:12.148691 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfr2r" event={"ID":"4fdae954-7222-4ebd-8b3a-205a9fc21094","Type":"ContainerStarted","Data":"fca47b7096e3a84e1eea1dc7f938923cc198bc09c083276c77effd5691f74260"} Dec 03 07:05:12 crc kubenswrapper[4818]: I1203 07:05:12.181207 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wfr2r" podStartSLOduration=2.5868269870000002 podStartE2EDuration="5.181165886s" podCreationTimestamp="2025-12-03 07:05:07 +0000 UTC" firstStartedPulling="2025-12-03 07:05:09.060855221 +0000 UTC m=+2266.752463973" lastFinishedPulling="2025-12-03 07:05:11.65519412 +0000 UTC m=+2269.346802872" observedRunningTime="2025-12-03 07:05:12.171100186 +0000 UTC m=+2269.862708978" watchObservedRunningTime="2025-12-03 07:05:12.181165886 +0000 UTC m=+2269.872774678" Dec 03 07:05:18 crc kubenswrapper[4818]: I1203 07:05:18.124751 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wfr2r" Dec 03 07:05:18 crc kubenswrapper[4818]: I1203 07:05:18.125211 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wfr2r" Dec 03 07:05:18 crc kubenswrapper[4818]: I1203 07:05:18.213973 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wfr2r" Dec 03 07:05:18 crc kubenswrapper[4818]: I1203 07:05:18.287656 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wfr2r" Dec 03 07:05:18 crc kubenswrapper[4818]: I1203 07:05:18.452645 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wfr2r"] Dec 03 07:05:20 crc kubenswrapper[4818]: I1203 07:05:20.262437 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wfr2r" podUID="4fdae954-7222-4ebd-8b3a-205a9fc21094" containerName="registry-server" containerID="cri-o://fca47b7096e3a84e1eea1dc7f938923cc198bc09c083276c77effd5691f74260" gracePeriod=2 Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.272941 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfr2r" Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.278778 4818 generic.go:334] "Generic (PLEG): container finished" podID="4fdae954-7222-4ebd-8b3a-205a9fc21094" containerID="fca47b7096e3a84e1eea1dc7f938923cc198bc09c083276c77effd5691f74260" exitCode=0 Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.278836 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfr2r" event={"ID":"4fdae954-7222-4ebd-8b3a-205a9fc21094","Type":"ContainerDied","Data":"fca47b7096e3a84e1eea1dc7f938923cc198bc09c083276c77effd5691f74260"} Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.278865 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wfr2r" event={"ID":"4fdae954-7222-4ebd-8b3a-205a9fc21094","Type":"ContainerDied","Data":"864bd65f590e2286349a6e1d7871795e2c7379dc374feb49d7cf28e57759776a"} Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.278882 4818 scope.go:117] "RemoveContainer" containerID="fca47b7096e3a84e1eea1dc7f938923cc198bc09c083276c77effd5691f74260" Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.279007 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wfr2r" Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.308603 4818 scope.go:117] "RemoveContainer" containerID="53f33a94a232e411ce813b77503be6fed91f737eb9b262a036ceb71eda3d99d1" Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.340108 4818 scope.go:117] "RemoveContainer" containerID="c9373927d9ab3148c6c8f0926d5f28a82e0d7e4663f087c6cfd2ebb64b9bfe7d" Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.379535 4818 scope.go:117] "RemoveContainer" containerID="fca47b7096e3a84e1eea1dc7f938923cc198bc09c083276c77effd5691f74260" Dec 03 07:05:21 crc kubenswrapper[4818]: E1203 07:05:21.380008 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fca47b7096e3a84e1eea1dc7f938923cc198bc09c083276c77effd5691f74260\": container with ID starting with fca47b7096e3a84e1eea1dc7f938923cc198bc09c083276c77effd5691f74260 not found: ID does not exist" containerID="fca47b7096e3a84e1eea1dc7f938923cc198bc09c083276c77effd5691f74260" Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.380067 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fca47b7096e3a84e1eea1dc7f938923cc198bc09c083276c77effd5691f74260"} err="failed to get container status \"fca47b7096e3a84e1eea1dc7f938923cc198bc09c083276c77effd5691f74260\": rpc error: code = NotFound desc = could not find container \"fca47b7096e3a84e1eea1dc7f938923cc198bc09c083276c77effd5691f74260\": container with ID starting with fca47b7096e3a84e1eea1dc7f938923cc198bc09c083276c77effd5691f74260 not found: ID does not exist" Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.380095 4818 scope.go:117] "RemoveContainer" containerID="53f33a94a232e411ce813b77503be6fed91f737eb9b262a036ceb71eda3d99d1" Dec 03 07:05:21 crc kubenswrapper[4818]: E1203 07:05:21.380462 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53f33a94a232e411ce813b77503be6fed91f737eb9b262a036ceb71eda3d99d1\": container with ID starting with 53f33a94a232e411ce813b77503be6fed91f737eb9b262a036ceb71eda3d99d1 not found: ID does not exist" containerID="53f33a94a232e411ce813b77503be6fed91f737eb9b262a036ceb71eda3d99d1" Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.380491 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53f33a94a232e411ce813b77503be6fed91f737eb9b262a036ceb71eda3d99d1"} err="failed to get container status \"53f33a94a232e411ce813b77503be6fed91f737eb9b262a036ceb71eda3d99d1\": rpc error: code = NotFound desc = could not find container \"53f33a94a232e411ce813b77503be6fed91f737eb9b262a036ceb71eda3d99d1\": container with ID starting with 53f33a94a232e411ce813b77503be6fed91f737eb9b262a036ceb71eda3d99d1 not found: ID does not exist" Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.380515 4818 scope.go:117] "RemoveContainer" containerID="c9373927d9ab3148c6c8f0926d5f28a82e0d7e4663f087c6cfd2ebb64b9bfe7d" Dec 03 07:05:21 crc kubenswrapper[4818]: E1203 07:05:21.380771 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9373927d9ab3148c6c8f0926d5f28a82e0d7e4663f087c6cfd2ebb64b9bfe7d\": container with ID starting with c9373927d9ab3148c6c8f0926d5f28a82e0d7e4663f087c6cfd2ebb64b9bfe7d not found: ID does not exist" containerID="c9373927d9ab3148c6c8f0926d5f28a82e0d7e4663f087c6cfd2ebb64b9bfe7d" Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.380801 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9373927d9ab3148c6c8f0926d5f28a82e0d7e4663f087c6cfd2ebb64b9bfe7d"} err="failed to get container status \"c9373927d9ab3148c6c8f0926d5f28a82e0d7e4663f087c6cfd2ebb64b9bfe7d\": rpc error: code = NotFound desc = could not find container \"c9373927d9ab3148c6c8f0926d5f28a82e0d7e4663f087c6cfd2ebb64b9bfe7d\": container with ID starting with c9373927d9ab3148c6c8f0926d5f28a82e0d7e4663f087c6cfd2ebb64b9bfe7d not found: ID does not exist" Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.417423 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fdae954-7222-4ebd-8b3a-205a9fc21094-catalog-content\") pod \"4fdae954-7222-4ebd-8b3a-205a9fc21094\" (UID: \"4fdae954-7222-4ebd-8b3a-205a9fc21094\") " Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.417501 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fdae954-7222-4ebd-8b3a-205a9fc21094-utilities\") pod \"4fdae954-7222-4ebd-8b3a-205a9fc21094\" (UID: \"4fdae954-7222-4ebd-8b3a-205a9fc21094\") " Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.417558 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbn9m\" (UniqueName: \"kubernetes.io/projected/4fdae954-7222-4ebd-8b3a-205a9fc21094-kube-api-access-cbn9m\") pod \"4fdae954-7222-4ebd-8b3a-205a9fc21094\" (UID: \"4fdae954-7222-4ebd-8b3a-205a9fc21094\") " Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.420626 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fdae954-7222-4ebd-8b3a-205a9fc21094-utilities" (OuterVolumeSpecName: "utilities") pod "4fdae954-7222-4ebd-8b3a-205a9fc21094" (UID: "4fdae954-7222-4ebd-8b3a-205a9fc21094"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.428690 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fdae954-7222-4ebd-8b3a-205a9fc21094-kube-api-access-cbn9m" (OuterVolumeSpecName: "kube-api-access-cbn9m") pod "4fdae954-7222-4ebd-8b3a-205a9fc21094" (UID: "4fdae954-7222-4ebd-8b3a-205a9fc21094"). InnerVolumeSpecName "kube-api-access-cbn9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.469672 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fdae954-7222-4ebd-8b3a-205a9fc21094-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4fdae954-7222-4ebd-8b3a-205a9fc21094" (UID: "4fdae954-7222-4ebd-8b3a-205a9fc21094"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.520558 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fdae954-7222-4ebd-8b3a-205a9fc21094-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.520593 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fdae954-7222-4ebd-8b3a-205a9fc21094-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.520621 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbn9m\" (UniqueName: \"kubernetes.io/projected/4fdae954-7222-4ebd-8b3a-205a9fc21094-kube-api-access-cbn9m\") on node \"crc\" DevicePath \"\"" Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.632343 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wfr2r"] Dec 03 07:05:21 crc kubenswrapper[4818]: I1203 07:05:21.642913 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wfr2r"] Dec 03 07:05:22 crc kubenswrapper[4818]: I1203 07:05:22.757598 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fdae954-7222-4ebd-8b3a-205a9fc21094" path="/var/lib/kubelet/pods/4fdae954-7222-4ebd-8b3a-205a9fc21094/volumes" Dec 03 07:05:23 crc kubenswrapper[4818]: I1203 07:05:23.738862 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:05:23 crc kubenswrapper[4818]: E1203 07:05:23.739536 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:05:36 crc kubenswrapper[4818]: I1203 07:05:36.738109 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:05:36 crc kubenswrapper[4818]: E1203 07:05:36.739225 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:05:51 crc kubenswrapper[4818]: I1203 07:05:51.737526 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:05:51 crc kubenswrapper[4818]: E1203 07:05:51.738594 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:06:05 crc kubenswrapper[4818]: I1203 07:06:05.737710 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:06:05 crc kubenswrapper[4818]: E1203 07:06:05.738515 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:06:20 crc kubenswrapper[4818]: I1203 07:06:20.737860 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:06:20 crc kubenswrapper[4818]: E1203 07:06:20.738670 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:06:32 crc kubenswrapper[4818]: I1203 07:06:32.747038 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:06:32 crc kubenswrapper[4818]: E1203 07:06:32.748069 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:06:46 crc kubenswrapper[4818]: I1203 07:06:46.738551 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:06:46 crc kubenswrapper[4818]: E1203 07:06:46.739449 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:06:58 crc kubenswrapper[4818]: I1203 07:06:58.738270 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:06:58 crc kubenswrapper[4818]: E1203 07:06:58.739120 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:07:12 crc kubenswrapper[4818]: I1203 07:07:12.748458 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:07:12 crc kubenswrapper[4818]: E1203 07:07:12.750083 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:07:25 crc kubenswrapper[4818]: I1203 07:07:25.738279 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:07:25 crc kubenswrapper[4818]: E1203 07:07:25.739493 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:07:37 crc kubenswrapper[4818]: I1203 07:07:37.737942 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:07:37 crc kubenswrapper[4818]: E1203 07:07:37.741695 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:07:52 crc kubenswrapper[4818]: I1203 07:07:52.745734 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:07:52 crc kubenswrapper[4818]: E1203 07:07:52.746639 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:08:03 crc kubenswrapper[4818]: I1203 07:08:03.738045 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:08:03 crc kubenswrapper[4818]: E1203 07:08:03.739089 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:08:04 crc kubenswrapper[4818]: E1203 07:08:04.431921 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5796328_5b74_4f5a_939b_403485be16a2.slice/crio-conmon-eb3936ef05c0b8dacf299ec5aab40e276a4c36802f2fc1fe76eef3624056eb68.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5796328_5b74_4f5a_939b_403485be16a2.slice/crio-eb3936ef05c0b8dacf299ec5aab40e276a4c36802f2fc1fe76eef3624056eb68.scope\": RecentStats: unable to find data in memory cache]" Dec 03 07:08:05 crc kubenswrapper[4818]: I1203 07:08:05.011003 4818 generic.go:334] "Generic (PLEG): container finished" podID="e5796328-5b74-4f5a-939b-403485be16a2" containerID="eb3936ef05c0b8dacf299ec5aab40e276a4c36802f2fc1fe76eef3624056eb68" exitCode=0 Dec 03 07:08:05 crc kubenswrapper[4818]: I1203 07:08:05.011124 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" event={"ID":"e5796328-5b74-4f5a-939b-403485be16a2","Type":"ContainerDied","Data":"eb3936ef05c0b8dacf299ec5aab40e276a4c36802f2fc1fe76eef3624056eb68"} Dec 03 07:08:06 crc kubenswrapper[4818]: I1203 07:08:06.586186 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:08:06 crc kubenswrapper[4818]: I1203 07:08:06.625245 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-ssh-key\") pod \"e5796328-5b74-4f5a-939b-403485be16a2\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " Dec 03 07:08:06 crc kubenswrapper[4818]: I1203 07:08:06.625300 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-libvirt-combined-ca-bundle\") pod \"e5796328-5b74-4f5a-939b-403485be16a2\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " Dec 03 07:08:06 crc kubenswrapper[4818]: I1203 07:08:06.625433 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7595\" (UniqueName: \"kubernetes.io/projected/e5796328-5b74-4f5a-939b-403485be16a2-kube-api-access-t7595\") pod \"e5796328-5b74-4f5a-939b-403485be16a2\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " Dec 03 07:08:06 crc kubenswrapper[4818]: I1203 07:08:06.625484 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-libvirt-secret-0\") pod \"e5796328-5b74-4f5a-939b-403485be16a2\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " Dec 03 07:08:06 crc kubenswrapper[4818]: I1203 07:08:06.625539 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-inventory\") pod \"e5796328-5b74-4f5a-939b-403485be16a2\" (UID: \"e5796328-5b74-4f5a-939b-403485be16a2\") " Dec 03 07:08:06 crc kubenswrapper[4818]: I1203 07:08:06.647986 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "e5796328-5b74-4f5a-939b-403485be16a2" (UID: "e5796328-5b74-4f5a-939b-403485be16a2"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:08:06 crc kubenswrapper[4818]: I1203 07:08:06.648397 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5796328-5b74-4f5a-939b-403485be16a2-kube-api-access-t7595" (OuterVolumeSpecName: "kube-api-access-t7595") pod "e5796328-5b74-4f5a-939b-403485be16a2" (UID: "e5796328-5b74-4f5a-939b-403485be16a2"). InnerVolumeSpecName "kube-api-access-t7595". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:08:06 crc kubenswrapper[4818]: I1203 07:08:06.659791 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-inventory" (OuterVolumeSpecName: "inventory") pod "e5796328-5b74-4f5a-939b-403485be16a2" (UID: "e5796328-5b74-4f5a-939b-403485be16a2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:08:06 crc kubenswrapper[4818]: I1203 07:08:06.661047 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "e5796328-5b74-4f5a-939b-403485be16a2" (UID: "e5796328-5b74-4f5a-939b-403485be16a2"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:08:06 crc kubenswrapper[4818]: I1203 07:08:06.668612 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e5796328-5b74-4f5a-939b-403485be16a2" (UID: "e5796328-5b74-4f5a-939b-403485be16a2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:08:06 crc kubenswrapper[4818]: I1203 07:08:06.727766 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 07:08:06 crc kubenswrapper[4818]: I1203 07:08:06.727799 4818 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:08:06 crc kubenswrapper[4818]: I1203 07:08:06.727830 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7595\" (UniqueName: \"kubernetes.io/projected/e5796328-5b74-4f5a-939b-403485be16a2-kube-api-access-t7595\") on node \"crc\" DevicePath \"\"" Dec 03 07:08:06 crc kubenswrapper[4818]: I1203 07:08:06.727840 4818 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:08:06 crc kubenswrapper[4818]: I1203 07:08:06.727848 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5796328-5b74-4f5a-939b-403485be16a2-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.037066 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" event={"ID":"e5796328-5b74-4f5a-939b-403485be16a2","Type":"ContainerDied","Data":"717ed273d52d5b2629e334549d3823db7956d5c389a1fb064ac2b19c3c53cf7f"} Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.037106 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="717ed273d52d5b2629e334549d3823db7956d5c389a1fb064ac2b19c3c53cf7f" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.037196 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.171774 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p"] Dec 03 07:08:07 crc kubenswrapper[4818]: E1203 07:08:07.172264 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5796328-5b74-4f5a-939b-403485be16a2" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.172286 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5796328-5b74-4f5a-939b-403485be16a2" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 07:08:07 crc kubenswrapper[4818]: E1203 07:08:07.172305 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fdae954-7222-4ebd-8b3a-205a9fc21094" containerName="extract-content" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.172314 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fdae954-7222-4ebd-8b3a-205a9fc21094" containerName="extract-content" Dec 03 07:08:07 crc kubenswrapper[4818]: E1203 07:08:07.172339 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fdae954-7222-4ebd-8b3a-205a9fc21094" containerName="registry-server" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.172348 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fdae954-7222-4ebd-8b3a-205a9fc21094" containerName="registry-server" Dec 03 07:08:07 crc kubenswrapper[4818]: E1203 07:08:07.172364 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fdae954-7222-4ebd-8b3a-205a9fc21094" containerName="extract-utilities" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.172372 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fdae954-7222-4ebd-8b3a-205a9fc21094" containerName="extract-utilities" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.172613 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5796328-5b74-4f5a-939b-403485be16a2" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.172643 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fdae954-7222-4ebd-8b3a-205a9fc21094" containerName="registry-server" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.173535 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.175730 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.176559 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.176974 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.177156 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.177281 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.177500 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.178376 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2bpsv" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.198516 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p"] Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.266409 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.266477 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.266525 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.266546 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzwgs\" (UniqueName: \"kubernetes.io/projected/febe4f5e-461e-4562-a168-b99a22ef6e7c-kube-api-access-dzwgs\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.266573 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.266594 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.266610 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.266630 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.266665 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.368399 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.368548 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.368602 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.368655 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.368681 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzwgs\" (UniqueName: \"kubernetes.io/projected/febe4f5e-461e-4562-a168-b99a22ef6e7c-kube-api-access-dzwgs\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.368713 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.368773 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.368800 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.368840 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.373808 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.374631 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.375327 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.375657 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.376337 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.377209 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.378577 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.381496 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.388971 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzwgs\" (UniqueName: \"kubernetes.io/projected/febe4f5e-461e-4562-a168-b99a22ef6e7c-kube-api-access-dzwgs\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4l2p\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:07 crc kubenswrapper[4818]: I1203 07:08:07.499975 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:08:08 crc kubenswrapper[4818]: I1203 07:08:08.117344 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p"] Dec 03 07:08:09 crc kubenswrapper[4818]: I1203 07:08:09.057719 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" event={"ID":"febe4f5e-461e-4562-a168-b99a22ef6e7c","Type":"ContainerStarted","Data":"99c73f6a02e879bcba3687d45a60345bea1ebecd99a685d8eccb6f5e53ba8247"} Dec 03 07:08:09 crc kubenswrapper[4818]: I1203 07:08:09.058077 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" event={"ID":"febe4f5e-461e-4562-a168-b99a22ef6e7c","Type":"ContainerStarted","Data":"d1fa27c39c61f5bf8dea2c6031cf43dc5d754f46a7378af8e188d8bb80397456"} Dec 03 07:08:17 crc kubenswrapper[4818]: I1203 07:08:17.737951 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:08:17 crc kubenswrapper[4818]: E1203 07:08:17.738929 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:08:28 crc kubenswrapper[4818]: I1203 07:08:28.738155 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:08:28 crc kubenswrapper[4818]: E1203 07:08:28.739300 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:08:40 crc kubenswrapper[4818]: I1203 07:08:40.738484 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:08:40 crc kubenswrapper[4818]: E1203 07:08:40.739276 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:08:55 crc kubenswrapper[4818]: I1203 07:08:55.737526 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:08:55 crc kubenswrapper[4818]: E1203 07:08:55.738168 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:09:09 crc kubenswrapper[4818]: I1203 07:09:09.738312 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:09:09 crc kubenswrapper[4818]: E1203 07:09:09.739583 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:09:22 crc kubenswrapper[4818]: I1203 07:09:22.743955 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:09:22 crc kubenswrapper[4818]: E1203 07:09:22.744640 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:09:35 crc kubenswrapper[4818]: I1203 07:09:35.738174 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:09:35 crc kubenswrapper[4818]: E1203 07:09:35.739156 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:09:47 crc kubenswrapper[4818]: I1203 07:09:47.738015 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:09:48 crc kubenswrapper[4818]: I1203 07:09:48.157384 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"ce526197baa6385b440d1405d4e8b2e96abba6ba7070564ccce6365e0b5b03e8"} Dec 03 07:09:48 crc kubenswrapper[4818]: I1203 07:09:48.182110 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" podStartSLOduration=100.708032873 podStartE2EDuration="1m41.182083569s" podCreationTimestamp="2025-12-03 07:08:07 +0000 UTC" firstStartedPulling="2025-12-03 07:08:08.12337578 +0000 UTC m=+2445.814984532" lastFinishedPulling="2025-12-03 07:08:08.597426436 +0000 UTC m=+2446.289035228" observedRunningTime="2025-12-03 07:08:09.083923682 +0000 UTC m=+2446.775532434" watchObservedRunningTime="2025-12-03 07:09:48.182083569 +0000 UTC m=+2545.873692331" Dec 03 07:11:24 crc kubenswrapper[4818]: I1203 07:11:24.256746 4818 generic.go:334] "Generic (PLEG): container finished" podID="febe4f5e-461e-4562-a168-b99a22ef6e7c" containerID="99c73f6a02e879bcba3687d45a60345bea1ebecd99a685d8eccb6f5e53ba8247" exitCode=0 Dec 03 07:11:24 crc kubenswrapper[4818]: I1203 07:11:24.257111 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" event={"ID":"febe4f5e-461e-4562-a168-b99a22ef6e7c","Type":"ContainerDied","Data":"99c73f6a02e879bcba3687d45a60345bea1ebecd99a685d8eccb6f5e53ba8247"} Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.703167 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.882789 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-migration-ssh-key-1\") pod \"febe4f5e-461e-4562-a168-b99a22ef6e7c\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.882925 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-cell1-compute-config-0\") pod \"febe4f5e-461e-4562-a168-b99a22ef6e7c\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.883015 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzwgs\" (UniqueName: \"kubernetes.io/projected/febe4f5e-461e-4562-a168-b99a22ef6e7c-kube-api-access-dzwgs\") pod \"febe4f5e-461e-4562-a168-b99a22ef6e7c\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.883044 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-migration-ssh-key-0\") pod \"febe4f5e-461e-4562-a168-b99a22ef6e7c\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.883118 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-combined-ca-bundle\") pod \"febe4f5e-461e-4562-a168-b99a22ef6e7c\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.883180 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-cell1-compute-config-1\") pod \"febe4f5e-461e-4562-a168-b99a22ef6e7c\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.883207 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-inventory\") pod \"febe4f5e-461e-4562-a168-b99a22ef6e7c\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.883238 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-extra-config-0\") pod \"febe4f5e-461e-4562-a168-b99a22ef6e7c\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.883265 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-ssh-key\") pod \"febe4f5e-461e-4562-a168-b99a22ef6e7c\" (UID: \"febe4f5e-461e-4562-a168-b99a22ef6e7c\") " Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.892366 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/febe4f5e-461e-4562-a168-b99a22ef6e7c-kube-api-access-dzwgs" (OuterVolumeSpecName: "kube-api-access-dzwgs") pod "febe4f5e-461e-4562-a168-b99a22ef6e7c" (UID: "febe4f5e-461e-4562-a168-b99a22ef6e7c"). InnerVolumeSpecName "kube-api-access-dzwgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.904133 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "febe4f5e-461e-4562-a168-b99a22ef6e7c" (UID: "febe4f5e-461e-4562-a168-b99a22ef6e7c"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.913648 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "febe4f5e-461e-4562-a168-b99a22ef6e7c" (UID: "febe4f5e-461e-4562-a168-b99a22ef6e7c"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.914546 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "febe4f5e-461e-4562-a168-b99a22ef6e7c" (UID: "febe4f5e-461e-4562-a168-b99a22ef6e7c"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.914656 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "febe4f5e-461e-4562-a168-b99a22ef6e7c" (UID: "febe4f5e-461e-4562-a168-b99a22ef6e7c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.922241 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-inventory" (OuterVolumeSpecName: "inventory") pod "febe4f5e-461e-4562-a168-b99a22ef6e7c" (UID: "febe4f5e-461e-4562-a168-b99a22ef6e7c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.922606 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "febe4f5e-461e-4562-a168-b99a22ef6e7c" (UID: "febe4f5e-461e-4562-a168-b99a22ef6e7c"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.927794 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "febe4f5e-461e-4562-a168-b99a22ef6e7c" (UID: "febe4f5e-461e-4562-a168-b99a22ef6e7c"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.957982 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "febe4f5e-461e-4562-a168-b99a22ef6e7c" (UID: "febe4f5e-461e-4562-a168-b99a22ef6e7c"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.984850 4818 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.984882 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzwgs\" (UniqueName: \"kubernetes.io/projected/febe4f5e-461e-4562-a168-b99a22ef6e7c-kube-api-access-dzwgs\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.984892 4818 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.984901 4818 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.984913 4818 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.984923 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.984932 4818 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.984940 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:25 crc kubenswrapper[4818]: I1203 07:11:25.984947 4818 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/febe4f5e-461e-4562-a168-b99a22ef6e7c-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.281807 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" event={"ID":"febe4f5e-461e-4562-a168-b99a22ef6e7c","Type":"ContainerDied","Data":"d1fa27c39c61f5bf8dea2c6031cf43dc5d754f46a7378af8e188d8bb80397456"} Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.281874 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1fa27c39c61f5bf8dea2c6031cf43dc5d754f46a7378af8e188d8bb80397456" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.281884 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4l2p" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.429178 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94"] Dec 03 07:11:26 crc kubenswrapper[4818]: E1203 07:11:26.430010 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="febe4f5e-461e-4562-a168-b99a22ef6e7c" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.430116 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="febe4f5e-461e-4562-a168-b99a22ef6e7c" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.430431 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="febe4f5e-461e-4562-a168-b99a22ef6e7c" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.431369 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.434429 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.434698 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2bpsv" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.434938 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.435247 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.435409 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.464508 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94"] Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.600847 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.600927 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.601083 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.601114 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdfjr\" (UniqueName: \"kubernetes.io/projected/9113c780-f00c-4f24-b000-26270c809028-kube-api-access-wdfjr\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.601180 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.601205 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.601252 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.703241 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.703361 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.703386 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.703429 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.703449 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdfjr\" (UniqueName: \"kubernetes.io/projected/9113c780-f00c-4f24-b000-26270c809028-kube-api-access-wdfjr\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.703498 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.703514 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.708362 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.708905 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.709163 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.709615 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.712269 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.719125 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.737404 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdfjr\" (UniqueName: \"kubernetes.io/projected/9113c780-f00c-4f24-b000-26270c809028-kube-api-access-wdfjr\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wkr94\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:26 crc kubenswrapper[4818]: I1203 07:11:26.769902 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:11:27 crc kubenswrapper[4818]: I1203 07:11:27.405584 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94"] Dec 03 07:11:27 crc kubenswrapper[4818]: I1203 07:11:27.414229 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 07:11:28 crc kubenswrapper[4818]: I1203 07:11:28.307216 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" event={"ID":"9113c780-f00c-4f24-b000-26270c809028","Type":"ContainerStarted","Data":"1778c32f3421fbacd4c2d06bea36e5b05dee2ec671453243c2979099bfe20d3f"} Dec 03 07:11:28 crc kubenswrapper[4818]: I1203 07:11:28.307565 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" event={"ID":"9113c780-f00c-4f24-b000-26270c809028","Type":"ContainerStarted","Data":"101bb3bd0b92db6d67d8bec179b9627e056421b6a10ed3d164cf9e9e63805941"} Dec 03 07:11:28 crc kubenswrapper[4818]: I1203 07:11:28.344169 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" podStartSLOduration=1.804234617 podStartE2EDuration="2.344118535s" podCreationTimestamp="2025-12-03 07:11:26 +0000 UTC" firstStartedPulling="2025-12-03 07:11:27.413883394 +0000 UTC m=+2645.105492166" lastFinishedPulling="2025-12-03 07:11:27.953767292 +0000 UTC m=+2645.645376084" observedRunningTime="2025-12-03 07:11:28.331710616 +0000 UTC m=+2646.023319388" watchObservedRunningTime="2025-12-03 07:11:28.344118535 +0000 UTC m=+2646.035727327" Dec 03 07:12:13 crc kubenswrapper[4818]: I1203 07:12:13.302557 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:12:13 crc kubenswrapper[4818]: I1203 07:12:13.303242 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:12:43 crc kubenswrapper[4818]: I1203 07:12:43.302726 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:12:43 crc kubenswrapper[4818]: I1203 07:12:43.303576 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:13:13 crc kubenswrapper[4818]: I1203 07:13:13.302625 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:13:13 crc kubenswrapper[4818]: I1203 07:13:13.303417 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:13:13 crc kubenswrapper[4818]: I1203 07:13:13.303472 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 07:13:13 crc kubenswrapper[4818]: I1203 07:13:13.304280 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ce526197baa6385b440d1405d4e8b2e96abba6ba7070564ccce6365e0b5b03e8"} pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:13:13 crc kubenswrapper[4818]: I1203 07:13:13.304367 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" containerID="cri-o://ce526197baa6385b440d1405d4e8b2e96abba6ba7070564ccce6365e0b5b03e8" gracePeriod=600 Dec 03 07:13:13 crc kubenswrapper[4818]: I1203 07:13:13.635722 4818 generic.go:334] "Generic (PLEG): container finished" podID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerID="ce526197baa6385b440d1405d4e8b2e96abba6ba7070564ccce6365e0b5b03e8" exitCode=0 Dec 03 07:13:13 crc kubenswrapper[4818]: I1203 07:13:13.635956 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerDied","Data":"ce526197baa6385b440d1405d4e8b2e96abba6ba7070564ccce6365e0b5b03e8"} Dec 03 07:13:13 crc kubenswrapper[4818]: I1203 07:13:13.636061 4818 scope.go:117] "RemoveContainer" containerID="46329d78c86c71eae867813e44d5a495650e45efef6acff353e0bd9a3d2d2806" Dec 03 07:13:14 crc kubenswrapper[4818]: I1203 07:13:14.650331 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0"} Dec 03 07:14:15 crc kubenswrapper[4818]: I1203 07:14:15.336409 4818 generic.go:334] "Generic (PLEG): container finished" podID="9113c780-f00c-4f24-b000-26270c809028" containerID="1778c32f3421fbacd4c2d06bea36e5b05dee2ec671453243c2979099bfe20d3f" exitCode=0 Dec 03 07:14:15 crc kubenswrapper[4818]: I1203 07:14:15.336516 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" event={"ID":"9113c780-f00c-4f24-b000-26270c809028","Type":"ContainerDied","Data":"1778c32f3421fbacd4c2d06bea36e5b05dee2ec671453243c2979099bfe20d3f"} Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.789098 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.860807 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-telemetry-combined-ca-bundle\") pod \"9113c780-f00c-4f24-b000-26270c809028\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.861025 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdfjr\" (UniqueName: \"kubernetes.io/projected/9113c780-f00c-4f24-b000-26270c809028-kube-api-access-wdfjr\") pod \"9113c780-f00c-4f24-b000-26270c809028\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.861142 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-inventory\") pod \"9113c780-f00c-4f24-b000-26270c809028\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.861175 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-0\") pod \"9113c780-f00c-4f24-b000-26270c809028\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.861207 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-1\") pod \"9113c780-f00c-4f24-b000-26270c809028\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.861237 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ssh-key\") pod \"9113c780-f00c-4f24-b000-26270c809028\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.861301 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-2\") pod \"9113c780-f00c-4f24-b000-26270c809028\" (UID: \"9113c780-f00c-4f24-b000-26270c809028\") " Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.868138 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9113c780-f00c-4f24-b000-26270c809028-kube-api-access-wdfjr" (OuterVolumeSpecName: "kube-api-access-wdfjr") pod "9113c780-f00c-4f24-b000-26270c809028" (UID: "9113c780-f00c-4f24-b000-26270c809028"). InnerVolumeSpecName "kube-api-access-wdfjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.868139 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "9113c780-f00c-4f24-b000-26270c809028" (UID: "9113c780-f00c-4f24-b000-26270c809028"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.891257 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "9113c780-f00c-4f24-b000-26270c809028" (UID: "9113c780-f00c-4f24-b000-26270c809028"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.891594 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "9113c780-f00c-4f24-b000-26270c809028" (UID: "9113c780-f00c-4f24-b000-26270c809028"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.892430 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9113c780-f00c-4f24-b000-26270c809028" (UID: "9113c780-f00c-4f24-b000-26270c809028"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.898151 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "9113c780-f00c-4f24-b000-26270c809028" (UID: "9113c780-f00c-4f24-b000-26270c809028"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.901888 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-inventory" (OuterVolumeSpecName: "inventory") pod "9113c780-f00c-4f24-b000-26270c809028" (UID: "9113c780-f00c-4f24-b000-26270c809028"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.963911 4818 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.963945 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.963965 4818 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.963984 4818 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.963998 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdfjr\" (UniqueName: \"kubernetes.io/projected/9113c780-f00c-4f24-b000-26270c809028-kube-api-access-wdfjr\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.964009 4818 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:16 crc kubenswrapper[4818]: I1203 07:14:16.964021 4818 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9113c780-f00c-4f24-b000-26270c809028-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 07:14:17 crc kubenswrapper[4818]: I1203 07:14:17.363288 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" event={"ID":"9113c780-f00c-4f24-b000-26270c809028","Type":"ContainerDied","Data":"101bb3bd0b92db6d67d8bec179b9627e056421b6a10ed3d164cf9e9e63805941"} Dec 03 07:14:17 crc kubenswrapper[4818]: I1203 07:14:17.363391 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="101bb3bd0b92db6d67d8bec179b9627e056421b6a10ed3d164cf9e9e63805941" Dec 03 07:14:17 crc kubenswrapper[4818]: I1203 07:14:17.363437 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wkr94" Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.156176 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv"] Dec 03 07:15:00 crc kubenswrapper[4818]: E1203 07:15:00.157526 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9113c780-f00c-4f24-b000-26270c809028" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.157551 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9113c780-f00c-4f24-b000-26270c809028" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.157985 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="9113c780-f00c-4f24-b000-26270c809028" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.159173 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv" Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.163138 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.166222 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.169749 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv"] Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.301982 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1272456f-4111-4618-b002-6c1b4496a20b-secret-volume\") pod \"collect-profiles-29412435-sbmcv\" (UID: \"1272456f-4111-4618-b002-6c1b4496a20b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv" Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.302159 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1272456f-4111-4618-b002-6c1b4496a20b-config-volume\") pod \"collect-profiles-29412435-sbmcv\" (UID: \"1272456f-4111-4618-b002-6c1b4496a20b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv" Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.302228 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btltb\" (UniqueName: \"kubernetes.io/projected/1272456f-4111-4618-b002-6c1b4496a20b-kube-api-access-btltb\") pod \"collect-profiles-29412435-sbmcv\" (UID: \"1272456f-4111-4618-b002-6c1b4496a20b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv" Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.403394 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btltb\" (UniqueName: \"kubernetes.io/projected/1272456f-4111-4618-b002-6c1b4496a20b-kube-api-access-btltb\") pod \"collect-profiles-29412435-sbmcv\" (UID: \"1272456f-4111-4618-b002-6c1b4496a20b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv" Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.403473 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1272456f-4111-4618-b002-6c1b4496a20b-secret-volume\") pod \"collect-profiles-29412435-sbmcv\" (UID: \"1272456f-4111-4618-b002-6c1b4496a20b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv" Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.403597 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1272456f-4111-4618-b002-6c1b4496a20b-config-volume\") pod \"collect-profiles-29412435-sbmcv\" (UID: \"1272456f-4111-4618-b002-6c1b4496a20b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv" Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.404594 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1272456f-4111-4618-b002-6c1b4496a20b-config-volume\") pod \"collect-profiles-29412435-sbmcv\" (UID: \"1272456f-4111-4618-b002-6c1b4496a20b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv" Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.414926 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1272456f-4111-4618-b002-6c1b4496a20b-secret-volume\") pod \"collect-profiles-29412435-sbmcv\" (UID: \"1272456f-4111-4618-b002-6c1b4496a20b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv" Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.437107 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btltb\" (UniqueName: \"kubernetes.io/projected/1272456f-4111-4618-b002-6c1b4496a20b-kube-api-access-btltb\") pod \"collect-profiles-29412435-sbmcv\" (UID: \"1272456f-4111-4618-b002-6c1b4496a20b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv" Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.523957 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv" Dec 03 07:15:00 crc kubenswrapper[4818]: I1203 07:15:00.969491 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv"] Dec 03 07:15:00 crc kubenswrapper[4818]: W1203 07:15:00.984797 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1272456f_4111_4618_b002_6c1b4496a20b.slice/crio-fcec6ac0925b2e9eadde132748b7a42a2291619bcec07efa290aa5f46c17774c WatchSource:0}: Error finding container fcec6ac0925b2e9eadde132748b7a42a2291619bcec07efa290aa5f46c17774c: Status 404 returned error can't find the container with id fcec6ac0925b2e9eadde132748b7a42a2291619bcec07efa290aa5f46c17774c Dec 03 07:15:01 crc kubenswrapper[4818]: I1203 07:15:01.107116 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv" event={"ID":"1272456f-4111-4618-b002-6c1b4496a20b","Type":"ContainerStarted","Data":"fcec6ac0925b2e9eadde132748b7a42a2291619bcec07efa290aa5f46c17774c"} Dec 03 07:15:02 crc kubenswrapper[4818]: I1203 07:15:02.122397 4818 generic.go:334] "Generic (PLEG): container finished" podID="1272456f-4111-4618-b002-6c1b4496a20b" containerID="19ff6c14dd7da8bc5f5024d9cfd9d2841c82add77c65b5e3eb2b59ee656c0acb" exitCode=0 Dec 03 07:15:02 crc kubenswrapper[4818]: I1203 07:15:02.122514 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv" event={"ID":"1272456f-4111-4618-b002-6c1b4496a20b","Type":"ContainerDied","Data":"19ff6c14dd7da8bc5f5024d9cfd9d2841c82add77c65b5e3eb2b59ee656c0acb"} Dec 03 07:15:03 crc kubenswrapper[4818]: I1203 07:15:03.598129 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv" Dec 03 07:15:03 crc kubenswrapper[4818]: I1203 07:15:03.772600 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1272456f-4111-4618-b002-6c1b4496a20b-config-volume\") pod \"1272456f-4111-4618-b002-6c1b4496a20b\" (UID: \"1272456f-4111-4618-b002-6c1b4496a20b\") " Dec 03 07:15:03 crc kubenswrapper[4818]: I1203 07:15:03.773016 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1272456f-4111-4618-b002-6c1b4496a20b-secret-volume\") pod \"1272456f-4111-4618-b002-6c1b4496a20b\" (UID: \"1272456f-4111-4618-b002-6c1b4496a20b\") " Dec 03 07:15:03 crc kubenswrapper[4818]: I1203 07:15:03.773316 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btltb\" (UniqueName: \"kubernetes.io/projected/1272456f-4111-4618-b002-6c1b4496a20b-kube-api-access-btltb\") pod \"1272456f-4111-4618-b002-6c1b4496a20b\" (UID: \"1272456f-4111-4618-b002-6c1b4496a20b\") " Dec 03 07:15:03 crc kubenswrapper[4818]: I1203 07:15:03.773519 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1272456f-4111-4618-b002-6c1b4496a20b-config-volume" (OuterVolumeSpecName: "config-volume") pod "1272456f-4111-4618-b002-6c1b4496a20b" (UID: "1272456f-4111-4618-b002-6c1b4496a20b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:15:03 crc kubenswrapper[4818]: I1203 07:15:03.774866 4818 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1272456f-4111-4618-b002-6c1b4496a20b-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:03 crc kubenswrapper[4818]: I1203 07:15:03.782019 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1272456f-4111-4618-b002-6c1b4496a20b-kube-api-access-btltb" (OuterVolumeSpecName: "kube-api-access-btltb") pod "1272456f-4111-4618-b002-6c1b4496a20b" (UID: "1272456f-4111-4618-b002-6c1b4496a20b"). InnerVolumeSpecName "kube-api-access-btltb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:15:03 crc kubenswrapper[4818]: I1203 07:15:03.789949 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1272456f-4111-4618-b002-6c1b4496a20b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1272456f-4111-4618-b002-6c1b4496a20b" (UID: "1272456f-4111-4618-b002-6c1b4496a20b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:15:03 crc kubenswrapper[4818]: I1203 07:15:03.877220 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btltb\" (UniqueName: \"kubernetes.io/projected/1272456f-4111-4618-b002-6c1b4496a20b-kube-api-access-btltb\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:03 crc kubenswrapper[4818]: I1203 07:15:03.877453 4818 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1272456f-4111-4618-b002-6c1b4496a20b-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:04 crc kubenswrapper[4818]: I1203 07:15:04.143081 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv" event={"ID":"1272456f-4111-4618-b002-6c1b4496a20b","Type":"ContainerDied","Data":"fcec6ac0925b2e9eadde132748b7a42a2291619bcec07efa290aa5f46c17774c"} Dec 03 07:15:04 crc kubenswrapper[4818]: I1203 07:15:04.143371 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcec6ac0925b2e9eadde132748b7a42a2291619bcec07efa290aa5f46c17774c" Dec 03 07:15:04 crc kubenswrapper[4818]: I1203 07:15:04.143143 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412435-sbmcv" Dec 03 07:15:04 crc kubenswrapper[4818]: I1203 07:15:04.680972 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9"] Dec 03 07:15:04 crc kubenswrapper[4818]: I1203 07:15:04.688773 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412390-c45b9"] Dec 03 07:15:04 crc kubenswrapper[4818]: I1203 07:15:04.750725 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f033be5e-5c71-4aeb-8825-bcf234d174a1" path="/var/lib/kubelet/pods/f033be5e-5c71-4aeb-8825-bcf234d174a1/volumes" Dec 03 07:15:07 crc kubenswrapper[4818]: I1203 07:15:07.809072 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hxvm5"] Dec 03 07:15:07 crc kubenswrapper[4818]: E1203 07:15:07.809925 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1272456f-4111-4618-b002-6c1b4496a20b" containerName="collect-profiles" Dec 03 07:15:07 crc kubenswrapper[4818]: I1203 07:15:07.809943 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="1272456f-4111-4618-b002-6c1b4496a20b" containerName="collect-profiles" Dec 03 07:15:07 crc kubenswrapper[4818]: I1203 07:15:07.810171 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="1272456f-4111-4618-b002-6c1b4496a20b" containerName="collect-profiles" Dec 03 07:15:07 crc kubenswrapper[4818]: I1203 07:15:07.812084 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hxvm5" Dec 03 07:15:07 crc kubenswrapper[4818]: I1203 07:15:07.828743 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hxvm5"] Dec 03 07:15:07 crc kubenswrapper[4818]: I1203 07:15:07.870336 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hxxr\" (UniqueName: \"kubernetes.io/projected/c7719288-56aa-4987-940d-7074b4163a71-kube-api-access-4hxxr\") pod \"certified-operators-hxvm5\" (UID: \"c7719288-56aa-4987-940d-7074b4163a71\") " pod="openshift-marketplace/certified-operators-hxvm5" Dec 03 07:15:07 crc kubenswrapper[4818]: I1203 07:15:07.870398 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7719288-56aa-4987-940d-7074b4163a71-utilities\") pod \"certified-operators-hxvm5\" (UID: \"c7719288-56aa-4987-940d-7074b4163a71\") " pod="openshift-marketplace/certified-operators-hxvm5" Dec 03 07:15:07 crc kubenswrapper[4818]: I1203 07:15:07.870454 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7719288-56aa-4987-940d-7074b4163a71-catalog-content\") pod \"certified-operators-hxvm5\" (UID: \"c7719288-56aa-4987-940d-7074b4163a71\") " pod="openshift-marketplace/certified-operators-hxvm5" Dec 03 07:15:07 crc kubenswrapper[4818]: I1203 07:15:07.972055 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7719288-56aa-4987-940d-7074b4163a71-utilities\") pod \"certified-operators-hxvm5\" (UID: \"c7719288-56aa-4987-940d-7074b4163a71\") " pod="openshift-marketplace/certified-operators-hxvm5" Dec 03 07:15:07 crc kubenswrapper[4818]: I1203 07:15:07.972110 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7719288-56aa-4987-940d-7074b4163a71-catalog-content\") pod \"certified-operators-hxvm5\" (UID: \"c7719288-56aa-4987-940d-7074b4163a71\") " pod="openshift-marketplace/certified-operators-hxvm5" Dec 03 07:15:07 crc kubenswrapper[4818]: I1203 07:15:07.972274 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hxxr\" (UniqueName: \"kubernetes.io/projected/c7719288-56aa-4987-940d-7074b4163a71-kube-api-access-4hxxr\") pod \"certified-operators-hxvm5\" (UID: \"c7719288-56aa-4987-940d-7074b4163a71\") " pod="openshift-marketplace/certified-operators-hxvm5" Dec 03 07:15:07 crc kubenswrapper[4818]: I1203 07:15:07.972926 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7719288-56aa-4987-940d-7074b4163a71-utilities\") pod \"certified-operators-hxvm5\" (UID: \"c7719288-56aa-4987-940d-7074b4163a71\") " pod="openshift-marketplace/certified-operators-hxvm5" Dec 03 07:15:07 crc kubenswrapper[4818]: I1203 07:15:07.972995 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7719288-56aa-4987-940d-7074b4163a71-catalog-content\") pod \"certified-operators-hxvm5\" (UID: \"c7719288-56aa-4987-940d-7074b4163a71\") " pod="openshift-marketplace/certified-operators-hxvm5" Dec 03 07:15:07 crc kubenswrapper[4818]: I1203 07:15:07.991183 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hxxr\" (UniqueName: \"kubernetes.io/projected/c7719288-56aa-4987-940d-7074b4163a71-kube-api-access-4hxxr\") pod \"certified-operators-hxvm5\" (UID: \"c7719288-56aa-4987-940d-7074b4163a71\") " pod="openshift-marketplace/certified-operators-hxvm5" Dec 03 07:15:08 crc kubenswrapper[4818]: I1203 07:15:08.141377 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hxvm5" Dec 03 07:15:08 crc kubenswrapper[4818]: I1203 07:15:08.678518 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hxvm5"] Dec 03 07:15:09 crc kubenswrapper[4818]: I1203 07:15:09.197804 4818 generic.go:334] "Generic (PLEG): container finished" podID="c7719288-56aa-4987-940d-7074b4163a71" containerID="4acd6aee7528c3fe99a00786bb698d15a1e3976c64303096322f918e2b84a24c" exitCode=0 Dec 03 07:15:09 crc kubenswrapper[4818]: I1203 07:15:09.197958 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hxvm5" event={"ID":"c7719288-56aa-4987-940d-7074b4163a71","Type":"ContainerDied","Data":"4acd6aee7528c3fe99a00786bb698d15a1e3976c64303096322f918e2b84a24c"} Dec 03 07:15:09 crc kubenswrapper[4818]: I1203 07:15:09.203184 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hxvm5" event={"ID":"c7719288-56aa-4987-940d-7074b4163a71","Type":"ContainerStarted","Data":"fd6dec3a69075757453df74245b2e9b68c43dbeb3cebf58ae9b6f04c708a4920"} Dec 03 07:15:10 crc kubenswrapper[4818]: I1203 07:15:10.189784 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tnps9"] Dec 03 07:15:10 crc kubenswrapper[4818]: I1203 07:15:10.192677 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tnps9" Dec 03 07:15:10 crc kubenswrapper[4818]: I1203 07:15:10.214698 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tnps9"] Dec 03 07:15:10 crc kubenswrapper[4818]: I1203 07:15:10.218692 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hxvm5" event={"ID":"c7719288-56aa-4987-940d-7074b4163a71","Type":"ContainerStarted","Data":"1baa5c9e1a1e6e1ee8607c9e03ee3a65f7afbcbbf7585ea8b99a20fbc8393fec"} Dec 03 07:15:10 crc kubenswrapper[4818]: I1203 07:15:10.222247 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgqhp\" (UniqueName: \"kubernetes.io/projected/3234c8dd-c10f-4c64-a885-ce072dc14d7a-kube-api-access-dgqhp\") pod \"redhat-operators-tnps9\" (UID: \"3234c8dd-c10f-4c64-a885-ce072dc14d7a\") " pod="openshift-marketplace/redhat-operators-tnps9" Dec 03 07:15:10 crc kubenswrapper[4818]: I1203 07:15:10.222372 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3234c8dd-c10f-4c64-a885-ce072dc14d7a-catalog-content\") pod \"redhat-operators-tnps9\" (UID: \"3234c8dd-c10f-4c64-a885-ce072dc14d7a\") " pod="openshift-marketplace/redhat-operators-tnps9" Dec 03 07:15:10 crc kubenswrapper[4818]: I1203 07:15:10.222443 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3234c8dd-c10f-4c64-a885-ce072dc14d7a-utilities\") pod \"redhat-operators-tnps9\" (UID: \"3234c8dd-c10f-4c64-a885-ce072dc14d7a\") " pod="openshift-marketplace/redhat-operators-tnps9" Dec 03 07:15:10 crc kubenswrapper[4818]: I1203 07:15:10.324309 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3234c8dd-c10f-4c64-a885-ce072dc14d7a-utilities\") pod \"redhat-operators-tnps9\" (UID: \"3234c8dd-c10f-4c64-a885-ce072dc14d7a\") " pod="openshift-marketplace/redhat-operators-tnps9" Dec 03 07:15:10 crc kubenswrapper[4818]: I1203 07:15:10.324513 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgqhp\" (UniqueName: \"kubernetes.io/projected/3234c8dd-c10f-4c64-a885-ce072dc14d7a-kube-api-access-dgqhp\") pod \"redhat-operators-tnps9\" (UID: \"3234c8dd-c10f-4c64-a885-ce072dc14d7a\") " pod="openshift-marketplace/redhat-operators-tnps9" Dec 03 07:15:10 crc kubenswrapper[4818]: I1203 07:15:10.324579 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3234c8dd-c10f-4c64-a885-ce072dc14d7a-catalog-content\") pod \"redhat-operators-tnps9\" (UID: \"3234c8dd-c10f-4c64-a885-ce072dc14d7a\") " pod="openshift-marketplace/redhat-operators-tnps9" Dec 03 07:15:10 crc kubenswrapper[4818]: I1203 07:15:10.325252 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3234c8dd-c10f-4c64-a885-ce072dc14d7a-catalog-content\") pod \"redhat-operators-tnps9\" (UID: \"3234c8dd-c10f-4c64-a885-ce072dc14d7a\") " pod="openshift-marketplace/redhat-operators-tnps9" Dec 03 07:15:10 crc kubenswrapper[4818]: I1203 07:15:10.325935 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3234c8dd-c10f-4c64-a885-ce072dc14d7a-utilities\") pod \"redhat-operators-tnps9\" (UID: \"3234c8dd-c10f-4c64-a885-ce072dc14d7a\") " pod="openshift-marketplace/redhat-operators-tnps9" Dec 03 07:15:10 crc kubenswrapper[4818]: I1203 07:15:10.350126 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgqhp\" (UniqueName: \"kubernetes.io/projected/3234c8dd-c10f-4c64-a885-ce072dc14d7a-kube-api-access-dgqhp\") pod \"redhat-operators-tnps9\" (UID: \"3234c8dd-c10f-4c64-a885-ce072dc14d7a\") " pod="openshift-marketplace/redhat-operators-tnps9" Dec 03 07:15:10 crc kubenswrapper[4818]: I1203 07:15:10.529264 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tnps9" Dec 03 07:15:11 crc kubenswrapper[4818]: I1203 07:15:11.040314 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tnps9"] Dec 03 07:15:11 crc kubenswrapper[4818]: W1203 07:15:11.040504 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3234c8dd_c10f_4c64_a885_ce072dc14d7a.slice/crio-63c4f00e71026f9af7dd9b9f0f2367eb8a9da37c04607414ec67a6d4421409ec WatchSource:0}: Error finding container 63c4f00e71026f9af7dd9b9f0f2367eb8a9da37c04607414ec67a6d4421409ec: Status 404 returned error can't find the container with id 63c4f00e71026f9af7dd9b9f0f2367eb8a9da37c04607414ec67a6d4421409ec Dec 03 07:15:11 crc kubenswrapper[4818]: I1203 07:15:11.226717 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnps9" event={"ID":"3234c8dd-c10f-4c64-a885-ce072dc14d7a","Type":"ContainerStarted","Data":"5dd0b1354a8abd4a745379710b206ab6fd0587df2e0dd085786626244609dee7"} Dec 03 07:15:11 crc kubenswrapper[4818]: I1203 07:15:11.227006 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnps9" event={"ID":"3234c8dd-c10f-4c64-a885-ce072dc14d7a","Type":"ContainerStarted","Data":"63c4f00e71026f9af7dd9b9f0f2367eb8a9da37c04607414ec67a6d4421409ec"} Dec 03 07:15:11 crc kubenswrapper[4818]: I1203 07:15:11.230174 4818 generic.go:334] "Generic (PLEG): container finished" podID="c7719288-56aa-4987-940d-7074b4163a71" containerID="1baa5c9e1a1e6e1ee8607c9e03ee3a65f7afbcbbf7585ea8b99a20fbc8393fec" exitCode=0 Dec 03 07:15:11 crc kubenswrapper[4818]: I1203 07:15:11.230224 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hxvm5" event={"ID":"c7719288-56aa-4987-940d-7074b4163a71","Type":"ContainerDied","Data":"1baa5c9e1a1e6e1ee8607c9e03ee3a65f7afbcbbf7585ea8b99a20fbc8393fec"} Dec 03 07:15:12 crc kubenswrapper[4818]: I1203 07:15:12.262342 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hxvm5" event={"ID":"c7719288-56aa-4987-940d-7074b4163a71","Type":"ContainerStarted","Data":"29441d83ac8c9948544389ca4aa15928ff51ac50d88cb93dd457ff44f6158d89"} Dec 03 07:15:12 crc kubenswrapper[4818]: I1203 07:15:12.266872 4818 generic.go:334] "Generic (PLEG): container finished" podID="3234c8dd-c10f-4c64-a885-ce072dc14d7a" containerID="5dd0b1354a8abd4a745379710b206ab6fd0587df2e0dd085786626244609dee7" exitCode=0 Dec 03 07:15:12 crc kubenswrapper[4818]: I1203 07:15:12.266935 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnps9" event={"ID":"3234c8dd-c10f-4c64-a885-ce072dc14d7a","Type":"ContainerDied","Data":"5dd0b1354a8abd4a745379710b206ab6fd0587df2e0dd085786626244609dee7"} Dec 03 07:15:12 crc kubenswrapper[4818]: I1203 07:15:12.287724 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hxvm5" podStartSLOduration=2.808982466 podStartE2EDuration="5.287684025s" podCreationTimestamp="2025-12-03 07:15:07 +0000 UTC" firstStartedPulling="2025-12-03 07:15:09.200706347 +0000 UTC m=+2866.892315189" lastFinishedPulling="2025-12-03 07:15:11.679407996 +0000 UTC m=+2869.371016748" observedRunningTime="2025-12-03 07:15:12.282449295 +0000 UTC m=+2869.974058077" watchObservedRunningTime="2025-12-03 07:15:12.287684025 +0000 UTC m=+2869.979292787" Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.279248 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnps9" event={"ID":"3234c8dd-c10f-4c64-a885-ce072dc14d7a","Type":"ContainerStarted","Data":"dcd57a1fde80dbdde9b654074db152b1892b0d09864ab034d6ec4d046ae0f661"} Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.302438 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.302514 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.834940 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.836237 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.842354 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.842466 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-rwqql" Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.842779 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.851028 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.877052 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.919035 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.919144 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.919171 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.919213 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxjcz\" (UniqueName: \"kubernetes.io/projected/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-kube-api-access-vxjcz\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.919236 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.919261 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-config-data\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.919300 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.919331 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:13 crc kubenswrapper[4818]: I1203 07:15:13.919379 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.021027 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.021084 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.021135 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxjcz\" (UniqueName: \"kubernetes.io/projected/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-kube-api-access-vxjcz\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.021193 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.021215 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-config-data\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.021259 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.021294 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.021349 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.021442 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.021737 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.021991 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.022077 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.022530 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.023424 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-config-data\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.027546 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.030404 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.036460 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.039175 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxjcz\" (UniqueName: \"kubernetes.io/projected/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-kube-api-access-vxjcz\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.053543 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.188743 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.317911 4818 generic.go:334] "Generic (PLEG): container finished" podID="3234c8dd-c10f-4c64-a885-ce072dc14d7a" containerID="dcd57a1fde80dbdde9b654074db152b1892b0d09864ab034d6ec4d046ae0f661" exitCode=0 Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.317960 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnps9" event={"ID":"3234c8dd-c10f-4c64-a885-ce072dc14d7a","Type":"ContainerDied","Data":"dcd57a1fde80dbdde9b654074db152b1892b0d09864ab034d6ec4d046ae0f661"} Dec 03 07:15:14 crc kubenswrapper[4818]: I1203 07:15:14.685601 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 07:15:15 crc kubenswrapper[4818]: I1203 07:15:15.336987 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnps9" event={"ID":"3234c8dd-c10f-4c64-a885-ce072dc14d7a","Type":"ContainerStarted","Data":"335671af33249ba61bb66680e8160164f6a97655a923f60fad74d61ab594f323"} Dec 03 07:15:15 crc kubenswrapper[4818]: I1203 07:15:15.340589 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf","Type":"ContainerStarted","Data":"4f25090c7c16f6d3a82b065182ed88abec3b200807a6c5b7588126d0ea5a1870"} Dec 03 07:15:15 crc kubenswrapper[4818]: I1203 07:15:15.361569 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tnps9" podStartSLOduration=2.860322788 podStartE2EDuration="5.361547536s" podCreationTimestamp="2025-12-03 07:15:10 +0000 UTC" firstStartedPulling="2025-12-03 07:15:12.269974644 +0000 UTC m=+2869.961583407" lastFinishedPulling="2025-12-03 07:15:14.771199403 +0000 UTC m=+2872.462808155" observedRunningTime="2025-12-03 07:15:15.361342751 +0000 UTC m=+2873.052951543" watchObservedRunningTime="2025-12-03 07:15:15.361547536 +0000 UTC m=+2873.053156298" Dec 03 07:15:18 crc kubenswrapper[4818]: I1203 07:15:18.142469 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hxvm5" Dec 03 07:15:18 crc kubenswrapper[4818]: I1203 07:15:18.143139 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hxvm5" Dec 03 07:15:18 crc kubenswrapper[4818]: I1203 07:15:18.197615 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hxvm5" Dec 03 07:15:18 crc kubenswrapper[4818]: I1203 07:15:18.425083 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hxvm5" Dec 03 07:15:19 crc kubenswrapper[4818]: I1203 07:15:19.380342 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hxvm5"] Dec 03 07:15:20 crc kubenswrapper[4818]: I1203 07:15:20.394482 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hxvm5" podUID="c7719288-56aa-4987-940d-7074b4163a71" containerName="registry-server" containerID="cri-o://29441d83ac8c9948544389ca4aa15928ff51ac50d88cb93dd457ff44f6158d89" gracePeriod=2 Dec 03 07:15:20 crc kubenswrapper[4818]: I1203 07:15:20.529652 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tnps9" Dec 03 07:15:20 crc kubenswrapper[4818]: I1203 07:15:20.529764 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tnps9" Dec 03 07:15:21 crc kubenswrapper[4818]: I1203 07:15:21.424127 4818 generic.go:334] "Generic (PLEG): container finished" podID="c7719288-56aa-4987-940d-7074b4163a71" containerID="29441d83ac8c9948544389ca4aa15928ff51ac50d88cb93dd457ff44f6158d89" exitCode=0 Dec 03 07:15:21 crc kubenswrapper[4818]: I1203 07:15:21.424543 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hxvm5" event={"ID":"c7719288-56aa-4987-940d-7074b4163a71","Type":"ContainerDied","Data":"29441d83ac8c9948544389ca4aa15928ff51ac50d88cb93dd457ff44f6158d89"} Dec 03 07:15:21 crc kubenswrapper[4818]: I1203 07:15:21.528520 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hxvm5" Dec 03 07:15:21 crc kubenswrapper[4818]: I1203 07:15:21.585530 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tnps9" podUID="3234c8dd-c10f-4c64-a885-ce072dc14d7a" containerName="registry-server" probeResult="failure" output=< Dec 03 07:15:21 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Dec 03 07:15:21 crc kubenswrapper[4818]: > Dec 03 07:15:21 crc kubenswrapper[4818]: I1203 07:15:21.714564 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7719288-56aa-4987-940d-7074b4163a71-utilities\") pod \"c7719288-56aa-4987-940d-7074b4163a71\" (UID: \"c7719288-56aa-4987-940d-7074b4163a71\") " Dec 03 07:15:21 crc kubenswrapper[4818]: I1203 07:15:21.714913 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hxxr\" (UniqueName: \"kubernetes.io/projected/c7719288-56aa-4987-940d-7074b4163a71-kube-api-access-4hxxr\") pod \"c7719288-56aa-4987-940d-7074b4163a71\" (UID: \"c7719288-56aa-4987-940d-7074b4163a71\") " Dec 03 07:15:21 crc kubenswrapper[4818]: I1203 07:15:21.715172 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7719288-56aa-4987-940d-7074b4163a71-catalog-content\") pod \"c7719288-56aa-4987-940d-7074b4163a71\" (UID: \"c7719288-56aa-4987-940d-7074b4163a71\") " Dec 03 07:15:21 crc kubenswrapper[4818]: I1203 07:15:21.715521 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7719288-56aa-4987-940d-7074b4163a71-utilities" (OuterVolumeSpecName: "utilities") pod "c7719288-56aa-4987-940d-7074b4163a71" (UID: "c7719288-56aa-4987-940d-7074b4163a71"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:15:21 crc kubenswrapper[4818]: I1203 07:15:21.716569 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7719288-56aa-4987-940d-7074b4163a71-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:21 crc kubenswrapper[4818]: I1203 07:15:21.733135 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7719288-56aa-4987-940d-7074b4163a71-kube-api-access-4hxxr" (OuterVolumeSpecName: "kube-api-access-4hxxr") pod "c7719288-56aa-4987-940d-7074b4163a71" (UID: "c7719288-56aa-4987-940d-7074b4163a71"). InnerVolumeSpecName "kube-api-access-4hxxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:15:21 crc kubenswrapper[4818]: I1203 07:15:21.765040 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7719288-56aa-4987-940d-7074b4163a71-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c7719288-56aa-4987-940d-7074b4163a71" (UID: "c7719288-56aa-4987-940d-7074b4163a71"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:15:21 crc kubenswrapper[4818]: I1203 07:15:21.821530 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hxxr\" (UniqueName: \"kubernetes.io/projected/c7719288-56aa-4987-940d-7074b4163a71-kube-api-access-4hxxr\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:21 crc kubenswrapper[4818]: I1203 07:15:21.821572 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7719288-56aa-4987-940d-7074b4163a71-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:22 crc kubenswrapper[4818]: I1203 07:15:22.435604 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hxvm5" event={"ID":"c7719288-56aa-4987-940d-7074b4163a71","Type":"ContainerDied","Data":"fd6dec3a69075757453df74245b2e9b68c43dbeb3cebf58ae9b6f04c708a4920"} Dec 03 07:15:22 crc kubenswrapper[4818]: I1203 07:15:22.435687 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hxvm5" Dec 03 07:15:22 crc kubenswrapper[4818]: I1203 07:15:22.435909 4818 scope.go:117] "RemoveContainer" containerID="29441d83ac8c9948544389ca4aa15928ff51ac50d88cb93dd457ff44f6158d89" Dec 03 07:15:22 crc kubenswrapper[4818]: I1203 07:15:22.468400 4818 scope.go:117] "RemoveContainer" containerID="1baa5c9e1a1e6e1ee8607c9e03ee3a65f7afbcbbf7585ea8b99a20fbc8393fec" Dec 03 07:15:22 crc kubenswrapper[4818]: I1203 07:15:22.493930 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hxvm5"] Dec 03 07:15:22 crc kubenswrapper[4818]: I1203 07:15:22.499381 4818 scope.go:117] "RemoveContainer" containerID="4acd6aee7528c3fe99a00786bb698d15a1e3976c64303096322f918e2b84a24c" Dec 03 07:15:22 crc kubenswrapper[4818]: I1203 07:15:22.504618 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hxvm5"] Dec 03 07:15:22 crc kubenswrapper[4818]: I1203 07:15:22.798968 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7719288-56aa-4987-940d-7074b4163a71" path="/var/lib/kubelet/pods/c7719288-56aa-4987-940d-7074b4163a71/volumes" Dec 03 07:15:30 crc kubenswrapper[4818]: I1203 07:15:30.592900 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tnps9" Dec 03 07:15:30 crc kubenswrapper[4818]: I1203 07:15:30.643872 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tnps9" Dec 03 07:15:30 crc kubenswrapper[4818]: I1203 07:15:30.830856 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tnps9"] Dec 03 07:15:32 crc kubenswrapper[4818]: I1203 07:15:32.551603 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tnps9" podUID="3234c8dd-c10f-4c64-a885-ce072dc14d7a" containerName="registry-server" containerID="cri-o://335671af33249ba61bb66680e8160164f6a97655a923f60fad74d61ab594f323" gracePeriod=2 Dec 03 07:15:33 crc kubenswrapper[4818]: I1203 07:15:33.561190 4818 generic.go:334] "Generic (PLEG): container finished" podID="3234c8dd-c10f-4c64-a885-ce072dc14d7a" containerID="335671af33249ba61bb66680e8160164f6a97655a923f60fad74d61ab594f323" exitCode=0 Dec 03 07:15:33 crc kubenswrapper[4818]: I1203 07:15:33.561291 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnps9" event={"ID":"3234c8dd-c10f-4c64-a885-ce072dc14d7a","Type":"ContainerDied","Data":"335671af33249ba61bb66680e8160164f6a97655a923f60fad74d61ab594f323"} Dec 03 07:15:40 crc kubenswrapper[4818]: E1203 07:15:40.530240 4818 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 335671af33249ba61bb66680e8160164f6a97655a923f60fad74d61ab594f323 is running failed: container process not found" containerID="335671af33249ba61bb66680e8160164f6a97655a923f60fad74d61ab594f323" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 07:15:40 crc kubenswrapper[4818]: E1203 07:15:40.531121 4818 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 335671af33249ba61bb66680e8160164f6a97655a923f60fad74d61ab594f323 is running failed: container process not found" containerID="335671af33249ba61bb66680e8160164f6a97655a923f60fad74d61ab594f323" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 07:15:40 crc kubenswrapper[4818]: E1203 07:15:40.531477 4818 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 335671af33249ba61bb66680e8160164f6a97655a923f60fad74d61ab594f323 is running failed: container process not found" containerID="335671af33249ba61bb66680e8160164f6a97655a923f60fad74d61ab594f323" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 07:15:40 crc kubenswrapper[4818]: E1203 07:15:40.531536 4818 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 335671af33249ba61bb66680e8160164f6a97655a923f60fad74d61ab594f323 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-tnps9" podUID="3234c8dd-c10f-4c64-a885-ce072dc14d7a" containerName="registry-server" Dec 03 07:15:43 crc kubenswrapper[4818]: I1203 07:15:43.302052 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:15:43 crc kubenswrapper[4818]: I1203 07:15:43.302512 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:15:48 crc kubenswrapper[4818]: E1203 07:15:48.881206 4818 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 03 07:15:48 crc kubenswrapper[4818]: E1203 07:15:48.882112 4818 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vxjcz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 07:15:48 crc kubenswrapper[4818]: E1203 07:15:48.883370 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf" Dec 03 07:15:49 crc kubenswrapper[4818]: I1203 07:15:49.173250 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tnps9" Dec 03 07:15:49 crc kubenswrapper[4818]: I1203 07:15:49.335441 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3234c8dd-c10f-4c64-a885-ce072dc14d7a-catalog-content\") pod \"3234c8dd-c10f-4c64-a885-ce072dc14d7a\" (UID: \"3234c8dd-c10f-4c64-a885-ce072dc14d7a\") " Dec 03 07:15:49 crc kubenswrapper[4818]: I1203 07:15:49.335501 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgqhp\" (UniqueName: \"kubernetes.io/projected/3234c8dd-c10f-4c64-a885-ce072dc14d7a-kube-api-access-dgqhp\") pod \"3234c8dd-c10f-4c64-a885-ce072dc14d7a\" (UID: \"3234c8dd-c10f-4c64-a885-ce072dc14d7a\") " Dec 03 07:15:49 crc kubenswrapper[4818]: I1203 07:15:49.335714 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3234c8dd-c10f-4c64-a885-ce072dc14d7a-utilities\") pod \"3234c8dd-c10f-4c64-a885-ce072dc14d7a\" (UID: \"3234c8dd-c10f-4c64-a885-ce072dc14d7a\") " Dec 03 07:15:49 crc kubenswrapper[4818]: I1203 07:15:49.336839 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3234c8dd-c10f-4c64-a885-ce072dc14d7a-utilities" (OuterVolumeSpecName: "utilities") pod "3234c8dd-c10f-4c64-a885-ce072dc14d7a" (UID: "3234c8dd-c10f-4c64-a885-ce072dc14d7a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:15:49 crc kubenswrapper[4818]: I1203 07:15:49.343288 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3234c8dd-c10f-4c64-a885-ce072dc14d7a-kube-api-access-dgqhp" (OuterVolumeSpecName: "kube-api-access-dgqhp") pod "3234c8dd-c10f-4c64-a885-ce072dc14d7a" (UID: "3234c8dd-c10f-4c64-a885-ce072dc14d7a"). InnerVolumeSpecName "kube-api-access-dgqhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:15:49 crc kubenswrapper[4818]: I1203 07:15:49.433477 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3234c8dd-c10f-4c64-a885-ce072dc14d7a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3234c8dd-c10f-4c64-a885-ce072dc14d7a" (UID: "3234c8dd-c10f-4c64-a885-ce072dc14d7a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:15:49 crc kubenswrapper[4818]: I1203 07:15:49.437547 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3234c8dd-c10f-4c64-a885-ce072dc14d7a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:49 crc kubenswrapper[4818]: I1203 07:15:49.437586 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgqhp\" (UniqueName: \"kubernetes.io/projected/3234c8dd-c10f-4c64-a885-ce072dc14d7a-kube-api-access-dgqhp\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:49 crc kubenswrapper[4818]: I1203 07:15:49.437603 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3234c8dd-c10f-4c64-a885-ce072dc14d7a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:15:49 crc kubenswrapper[4818]: I1203 07:15:49.726697 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tnps9" Dec 03 07:15:49 crc kubenswrapper[4818]: I1203 07:15:49.726666 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tnps9" event={"ID":"3234c8dd-c10f-4c64-a885-ce072dc14d7a","Type":"ContainerDied","Data":"63c4f00e71026f9af7dd9b9f0f2367eb8a9da37c04607414ec67a6d4421409ec"} Dec 03 07:15:49 crc kubenswrapper[4818]: I1203 07:15:49.726905 4818 scope.go:117] "RemoveContainer" containerID="335671af33249ba61bb66680e8160164f6a97655a923f60fad74d61ab594f323" Dec 03 07:15:49 crc kubenswrapper[4818]: E1203 07:15:49.729454 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf" Dec 03 07:15:49 crc kubenswrapper[4818]: I1203 07:15:49.768191 4818 scope.go:117] "RemoveContainer" containerID="dcd57a1fde80dbdde9b654074db152b1892b0d09864ab034d6ec4d046ae0f661" Dec 03 07:15:49 crc kubenswrapper[4818]: I1203 07:15:49.826442 4818 scope.go:117] "RemoveContainer" containerID="5dd0b1354a8abd4a745379710b206ab6fd0587df2e0dd085786626244609dee7" Dec 03 07:15:49 crc kubenswrapper[4818]: I1203 07:15:49.842605 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tnps9"] Dec 03 07:15:49 crc kubenswrapper[4818]: I1203 07:15:49.854210 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tnps9"] Dec 03 07:15:50 crc kubenswrapper[4818]: I1203 07:15:50.754639 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3234c8dd-c10f-4c64-a885-ce072dc14d7a" path="/var/lib/kubelet/pods/3234c8dd-c10f-4c64-a885-ce072dc14d7a/volumes" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.287513 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4nq25"] Dec 03 07:15:52 crc kubenswrapper[4818]: E1203 07:15:52.288309 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3234c8dd-c10f-4c64-a885-ce072dc14d7a" containerName="extract-utilities" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.288328 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="3234c8dd-c10f-4c64-a885-ce072dc14d7a" containerName="extract-utilities" Dec 03 07:15:52 crc kubenswrapper[4818]: E1203 07:15:52.288347 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7719288-56aa-4987-940d-7074b4163a71" containerName="registry-server" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.288355 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7719288-56aa-4987-940d-7074b4163a71" containerName="registry-server" Dec 03 07:15:52 crc kubenswrapper[4818]: E1203 07:15:52.288372 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7719288-56aa-4987-940d-7074b4163a71" containerName="extract-utilities" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.288380 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7719288-56aa-4987-940d-7074b4163a71" containerName="extract-utilities" Dec 03 07:15:52 crc kubenswrapper[4818]: E1203 07:15:52.288394 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3234c8dd-c10f-4c64-a885-ce072dc14d7a" containerName="registry-server" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.288401 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="3234c8dd-c10f-4c64-a885-ce072dc14d7a" containerName="registry-server" Dec 03 07:15:52 crc kubenswrapper[4818]: E1203 07:15:52.288424 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3234c8dd-c10f-4c64-a885-ce072dc14d7a" containerName="extract-content" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.288432 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="3234c8dd-c10f-4c64-a885-ce072dc14d7a" containerName="extract-content" Dec 03 07:15:52 crc kubenswrapper[4818]: E1203 07:15:52.288449 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7719288-56aa-4987-940d-7074b4163a71" containerName="extract-content" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.288457 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7719288-56aa-4987-940d-7074b4163a71" containerName="extract-content" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.288712 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="3234c8dd-c10f-4c64-a885-ce072dc14d7a" containerName="registry-server" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.288742 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7719288-56aa-4987-940d-7074b4163a71" containerName="registry-server" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.299621 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4nq25" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.351901 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4nq25"] Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.425389 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32413695-c405-402e-bc2b-1eda396326f7-utilities\") pod \"community-operators-4nq25\" (UID: \"32413695-c405-402e-bc2b-1eda396326f7\") " pod="openshift-marketplace/community-operators-4nq25" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.425500 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32413695-c405-402e-bc2b-1eda396326f7-catalog-content\") pod \"community-operators-4nq25\" (UID: \"32413695-c405-402e-bc2b-1eda396326f7\") " pod="openshift-marketplace/community-operators-4nq25" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.425572 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxqps\" (UniqueName: \"kubernetes.io/projected/32413695-c405-402e-bc2b-1eda396326f7-kube-api-access-xxqps\") pod \"community-operators-4nq25\" (UID: \"32413695-c405-402e-bc2b-1eda396326f7\") " pod="openshift-marketplace/community-operators-4nq25" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.528158 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32413695-c405-402e-bc2b-1eda396326f7-catalog-content\") pod \"community-operators-4nq25\" (UID: \"32413695-c405-402e-bc2b-1eda396326f7\") " pod="openshift-marketplace/community-operators-4nq25" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.528290 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxqps\" (UniqueName: \"kubernetes.io/projected/32413695-c405-402e-bc2b-1eda396326f7-kube-api-access-xxqps\") pod \"community-operators-4nq25\" (UID: \"32413695-c405-402e-bc2b-1eda396326f7\") " pod="openshift-marketplace/community-operators-4nq25" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.528388 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32413695-c405-402e-bc2b-1eda396326f7-utilities\") pod \"community-operators-4nq25\" (UID: \"32413695-c405-402e-bc2b-1eda396326f7\") " pod="openshift-marketplace/community-operators-4nq25" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.529098 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32413695-c405-402e-bc2b-1eda396326f7-utilities\") pod \"community-operators-4nq25\" (UID: \"32413695-c405-402e-bc2b-1eda396326f7\") " pod="openshift-marketplace/community-operators-4nq25" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.529388 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32413695-c405-402e-bc2b-1eda396326f7-catalog-content\") pod \"community-operators-4nq25\" (UID: \"32413695-c405-402e-bc2b-1eda396326f7\") " pod="openshift-marketplace/community-operators-4nq25" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.558081 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxqps\" (UniqueName: \"kubernetes.io/projected/32413695-c405-402e-bc2b-1eda396326f7-kube-api-access-xxqps\") pod \"community-operators-4nq25\" (UID: \"32413695-c405-402e-bc2b-1eda396326f7\") " pod="openshift-marketplace/community-operators-4nq25" Dec 03 07:15:52 crc kubenswrapper[4818]: I1203 07:15:52.641966 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4nq25" Dec 03 07:15:53 crc kubenswrapper[4818]: I1203 07:15:53.190622 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4nq25"] Dec 03 07:15:53 crc kubenswrapper[4818]: I1203 07:15:53.775704 4818 generic.go:334] "Generic (PLEG): container finished" podID="32413695-c405-402e-bc2b-1eda396326f7" containerID="6e79767141b4a8a622a27dab0bd28f44023ad278f6939fa2f0cd14723979aba5" exitCode=0 Dec 03 07:15:53 crc kubenswrapper[4818]: I1203 07:15:53.775758 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4nq25" event={"ID":"32413695-c405-402e-bc2b-1eda396326f7","Type":"ContainerDied","Data":"6e79767141b4a8a622a27dab0bd28f44023ad278f6939fa2f0cd14723979aba5"} Dec 03 07:15:53 crc kubenswrapper[4818]: I1203 07:15:53.775789 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4nq25" event={"ID":"32413695-c405-402e-bc2b-1eda396326f7","Type":"ContainerStarted","Data":"3294185b6b3f848752fe79f54e82c2c2c7f8b10f5ac26e8b1fce2c0ec6310b81"} Dec 03 07:15:54 crc kubenswrapper[4818]: I1203 07:15:54.787838 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4nq25" event={"ID":"32413695-c405-402e-bc2b-1eda396326f7","Type":"ContainerStarted","Data":"3fe27c868b87f0bf3feb522aac5fab55a42c6a38df855551913c764b845c70b6"} Dec 03 07:15:55 crc kubenswrapper[4818]: I1203 07:15:55.798011 4818 generic.go:334] "Generic (PLEG): container finished" podID="32413695-c405-402e-bc2b-1eda396326f7" containerID="3fe27c868b87f0bf3feb522aac5fab55a42c6a38df855551913c764b845c70b6" exitCode=0 Dec 03 07:15:55 crc kubenswrapper[4818]: I1203 07:15:55.798060 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4nq25" event={"ID":"32413695-c405-402e-bc2b-1eda396326f7","Type":"ContainerDied","Data":"3fe27c868b87f0bf3feb522aac5fab55a42c6a38df855551913c764b845c70b6"} Dec 03 07:15:56 crc kubenswrapper[4818]: I1203 07:15:56.279836 4818 scope.go:117] "RemoveContainer" containerID="820336f7aa4564a8c9d5561a937f63298b3615bc679942f01a71a37b9a3328be" Dec 03 07:15:56 crc kubenswrapper[4818]: I1203 07:15:56.809990 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4nq25" event={"ID":"32413695-c405-402e-bc2b-1eda396326f7","Type":"ContainerStarted","Data":"58507d47202be00d190cdff97046c67d7ecd902701fe51c08f391acecb809505"} Dec 03 07:15:56 crc kubenswrapper[4818]: I1203 07:15:56.845835 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4nq25" podStartSLOduration=2.396528321 podStartE2EDuration="4.845774668s" podCreationTimestamp="2025-12-03 07:15:52 +0000 UTC" firstStartedPulling="2025-12-03 07:15:53.779371191 +0000 UTC m=+2911.470979943" lastFinishedPulling="2025-12-03 07:15:56.228617538 +0000 UTC m=+2913.920226290" observedRunningTime="2025-12-03 07:15:56.836156759 +0000 UTC m=+2914.527765531" watchObservedRunningTime="2025-12-03 07:15:56.845774668 +0000 UTC m=+2914.537383460" Dec 03 07:16:02 crc kubenswrapper[4818]: I1203 07:16:02.589912 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 03 07:16:02 crc kubenswrapper[4818]: I1203 07:16:02.644197 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4nq25" Dec 03 07:16:02 crc kubenswrapper[4818]: I1203 07:16:02.644252 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4nq25" Dec 03 07:16:02 crc kubenswrapper[4818]: I1203 07:16:02.722347 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4nq25" Dec 03 07:16:02 crc kubenswrapper[4818]: I1203 07:16:02.969626 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4nq25" Dec 03 07:16:03 crc kubenswrapper[4818]: I1203 07:16:03.018563 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4nq25"] Dec 03 07:16:03 crc kubenswrapper[4818]: I1203 07:16:03.930004 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf","Type":"ContainerStarted","Data":"cd6cbaba39d8cfd15e18267f59419e26bc7d2c41b6b936f55775ba3ebddd71c4"} Dec 03 07:16:03 crc kubenswrapper[4818]: I1203 07:16:03.958945 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.064099107 podStartE2EDuration="51.958915277s" podCreationTimestamp="2025-12-03 07:15:12 +0000 UTC" firstStartedPulling="2025-12-03 07:15:14.690918347 +0000 UTC m=+2872.382527139" lastFinishedPulling="2025-12-03 07:16:02.585734557 +0000 UTC m=+2920.277343309" observedRunningTime="2025-12-03 07:16:03.951024171 +0000 UTC m=+2921.642632963" watchObservedRunningTime="2025-12-03 07:16:03.958915277 +0000 UTC m=+2921.650524059" Dec 03 07:16:04 crc kubenswrapper[4818]: I1203 07:16:04.944465 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4nq25" podUID="32413695-c405-402e-bc2b-1eda396326f7" containerName="registry-server" containerID="cri-o://58507d47202be00d190cdff97046c67d7ecd902701fe51c08f391acecb809505" gracePeriod=2 Dec 03 07:16:05 crc kubenswrapper[4818]: I1203 07:16:05.469990 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4nq25" Dec 03 07:16:05 crc kubenswrapper[4818]: I1203 07:16:05.532475 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32413695-c405-402e-bc2b-1eda396326f7-catalog-content\") pod \"32413695-c405-402e-bc2b-1eda396326f7\" (UID: \"32413695-c405-402e-bc2b-1eda396326f7\") " Dec 03 07:16:05 crc kubenswrapper[4818]: I1203 07:16:05.532758 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32413695-c405-402e-bc2b-1eda396326f7-utilities\") pod \"32413695-c405-402e-bc2b-1eda396326f7\" (UID: \"32413695-c405-402e-bc2b-1eda396326f7\") " Dec 03 07:16:05 crc kubenswrapper[4818]: I1203 07:16:05.532837 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxqps\" (UniqueName: \"kubernetes.io/projected/32413695-c405-402e-bc2b-1eda396326f7-kube-api-access-xxqps\") pod \"32413695-c405-402e-bc2b-1eda396326f7\" (UID: \"32413695-c405-402e-bc2b-1eda396326f7\") " Dec 03 07:16:05 crc kubenswrapper[4818]: I1203 07:16:05.534083 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32413695-c405-402e-bc2b-1eda396326f7-utilities" (OuterVolumeSpecName: "utilities") pod "32413695-c405-402e-bc2b-1eda396326f7" (UID: "32413695-c405-402e-bc2b-1eda396326f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:05 crc kubenswrapper[4818]: I1203 07:16:05.541077 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32413695-c405-402e-bc2b-1eda396326f7-kube-api-access-xxqps" (OuterVolumeSpecName: "kube-api-access-xxqps") pod "32413695-c405-402e-bc2b-1eda396326f7" (UID: "32413695-c405-402e-bc2b-1eda396326f7"). InnerVolumeSpecName "kube-api-access-xxqps". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:16:05 crc kubenswrapper[4818]: I1203 07:16:05.637621 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxqps\" (UniqueName: \"kubernetes.io/projected/32413695-c405-402e-bc2b-1eda396326f7-kube-api-access-xxqps\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4818]: I1203 07:16:05.637654 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32413695-c405-402e-bc2b-1eda396326f7-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4818]: I1203 07:16:05.720101 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32413695-c405-402e-bc2b-1eda396326f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "32413695-c405-402e-bc2b-1eda396326f7" (UID: "32413695-c405-402e-bc2b-1eda396326f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:16:05 crc kubenswrapper[4818]: I1203 07:16:05.741806 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32413695-c405-402e-bc2b-1eda396326f7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:16:05 crc kubenswrapper[4818]: I1203 07:16:05.954981 4818 generic.go:334] "Generic (PLEG): container finished" podID="32413695-c405-402e-bc2b-1eda396326f7" containerID="58507d47202be00d190cdff97046c67d7ecd902701fe51c08f391acecb809505" exitCode=0 Dec 03 07:16:05 crc kubenswrapper[4818]: I1203 07:16:05.955035 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4nq25" event={"ID":"32413695-c405-402e-bc2b-1eda396326f7","Type":"ContainerDied","Data":"58507d47202be00d190cdff97046c67d7ecd902701fe51c08f391acecb809505"} Dec 03 07:16:05 crc kubenswrapper[4818]: I1203 07:16:05.955067 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4nq25" event={"ID":"32413695-c405-402e-bc2b-1eda396326f7","Type":"ContainerDied","Data":"3294185b6b3f848752fe79f54e82c2c2c7f8b10f5ac26e8b1fce2c0ec6310b81"} Dec 03 07:16:05 crc kubenswrapper[4818]: I1203 07:16:05.955088 4818 scope.go:117] "RemoveContainer" containerID="58507d47202be00d190cdff97046c67d7ecd902701fe51c08f391acecb809505" Dec 03 07:16:05 crc kubenswrapper[4818]: I1203 07:16:05.955254 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4nq25" Dec 03 07:16:06 crc kubenswrapper[4818]: I1203 07:16:06.003879 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4nq25"] Dec 03 07:16:06 crc kubenswrapper[4818]: I1203 07:16:06.011830 4818 scope.go:117] "RemoveContainer" containerID="3fe27c868b87f0bf3feb522aac5fab55a42c6a38df855551913c764b845c70b6" Dec 03 07:16:06 crc kubenswrapper[4818]: I1203 07:16:06.017329 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4nq25"] Dec 03 07:16:06 crc kubenswrapper[4818]: I1203 07:16:06.038013 4818 scope.go:117] "RemoveContainer" containerID="6e79767141b4a8a622a27dab0bd28f44023ad278f6939fa2f0cd14723979aba5" Dec 03 07:16:06 crc kubenswrapper[4818]: I1203 07:16:06.080610 4818 scope.go:117] "RemoveContainer" containerID="58507d47202be00d190cdff97046c67d7ecd902701fe51c08f391acecb809505" Dec 03 07:16:06 crc kubenswrapper[4818]: E1203 07:16:06.081128 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58507d47202be00d190cdff97046c67d7ecd902701fe51c08f391acecb809505\": container with ID starting with 58507d47202be00d190cdff97046c67d7ecd902701fe51c08f391acecb809505 not found: ID does not exist" containerID="58507d47202be00d190cdff97046c67d7ecd902701fe51c08f391acecb809505" Dec 03 07:16:06 crc kubenswrapper[4818]: I1203 07:16:06.081159 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58507d47202be00d190cdff97046c67d7ecd902701fe51c08f391acecb809505"} err="failed to get container status \"58507d47202be00d190cdff97046c67d7ecd902701fe51c08f391acecb809505\": rpc error: code = NotFound desc = could not find container \"58507d47202be00d190cdff97046c67d7ecd902701fe51c08f391acecb809505\": container with ID starting with 58507d47202be00d190cdff97046c67d7ecd902701fe51c08f391acecb809505 not found: ID does not exist" Dec 03 07:16:06 crc kubenswrapper[4818]: I1203 07:16:06.081184 4818 scope.go:117] "RemoveContainer" containerID="3fe27c868b87f0bf3feb522aac5fab55a42c6a38df855551913c764b845c70b6" Dec 03 07:16:06 crc kubenswrapper[4818]: E1203 07:16:06.081423 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fe27c868b87f0bf3feb522aac5fab55a42c6a38df855551913c764b845c70b6\": container with ID starting with 3fe27c868b87f0bf3feb522aac5fab55a42c6a38df855551913c764b845c70b6 not found: ID does not exist" containerID="3fe27c868b87f0bf3feb522aac5fab55a42c6a38df855551913c764b845c70b6" Dec 03 07:16:06 crc kubenswrapper[4818]: I1203 07:16:06.081444 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fe27c868b87f0bf3feb522aac5fab55a42c6a38df855551913c764b845c70b6"} err="failed to get container status \"3fe27c868b87f0bf3feb522aac5fab55a42c6a38df855551913c764b845c70b6\": rpc error: code = NotFound desc = could not find container \"3fe27c868b87f0bf3feb522aac5fab55a42c6a38df855551913c764b845c70b6\": container with ID starting with 3fe27c868b87f0bf3feb522aac5fab55a42c6a38df855551913c764b845c70b6 not found: ID does not exist" Dec 03 07:16:06 crc kubenswrapper[4818]: I1203 07:16:06.081457 4818 scope.go:117] "RemoveContainer" containerID="6e79767141b4a8a622a27dab0bd28f44023ad278f6939fa2f0cd14723979aba5" Dec 03 07:16:06 crc kubenswrapper[4818]: E1203 07:16:06.081662 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e79767141b4a8a622a27dab0bd28f44023ad278f6939fa2f0cd14723979aba5\": container with ID starting with 6e79767141b4a8a622a27dab0bd28f44023ad278f6939fa2f0cd14723979aba5 not found: ID does not exist" containerID="6e79767141b4a8a622a27dab0bd28f44023ad278f6939fa2f0cd14723979aba5" Dec 03 07:16:06 crc kubenswrapper[4818]: I1203 07:16:06.081681 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e79767141b4a8a622a27dab0bd28f44023ad278f6939fa2f0cd14723979aba5"} err="failed to get container status \"6e79767141b4a8a622a27dab0bd28f44023ad278f6939fa2f0cd14723979aba5\": rpc error: code = NotFound desc = could not find container \"6e79767141b4a8a622a27dab0bd28f44023ad278f6939fa2f0cd14723979aba5\": container with ID starting with 6e79767141b4a8a622a27dab0bd28f44023ad278f6939fa2f0cd14723979aba5 not found: ID does not exist" Dec 03 07:16:06 crc kubenswrapper[4818]: I1203 07:16:06.771120 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32413695-c405-402e-bc2b-1eda396326f7" path="/var/lib/kubelet/pods/32413695-c405-402e-bc2b-1eda396326f7/volumes" Dec 03 07:16:13 crc kubenswrapper[4818]: I1203 07:16:13.302869 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:16:13 crc kubenswrapper[4818]: I1203 07:16:13.303429 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:16:13 crc kubenswrapper[4818]: I1203 07:16:13.303481 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 07:16:13 crc kubenswrapper[4818]: I1203 07:16:13.304162 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0"} pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:16:13 crc kubenswrapper[4818]: I1203 07:16:13.304236 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" containerID="cri-o://9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" gracePeriod=600 Dec 03 07:16:13 crc kubenswrapper[4818]: E1203 07:16:13.434450 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:16:14 crc kubenswrapper[4818]: I1203 07:16:14.046799 4818 generic.go:334] "Generic (PLEG): container finished" podID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" exitCode=0 Dec 03 07:16:14 crc kubenswrapper[4818]: I1203 07:16:14.046854 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerDied","Data":"9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0"} Dec 03 07:16:14 crc kubenswrapper[4818]: I1203 07:16:14.047110 4818 scope.go:117] "RemoveContainer" containerID="ce526197baa6385b440d1405d4e8b2e96abba6ba7070564ccce6365e0b5b03e8" Dec 03 07:16:14 crc kubenswrapper[4818]: I1203 07:16:14.047894 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:16:14 crc kubenswrapper[4818]: E1203 07:16:14.048257 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:16:25 crc kubenswrapper[4818]: I1203 07:16:25.738514 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:16:25 crc kubenswrapper[4818]: E1203 07:16:25.739680 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:16:38 crc kubenswrapper[4818]: I1203 07:16:38.737617 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:16:38 crc kubenswrapper[4818]: E1203 07:16:38.738511 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:16:50 crc kubenswrapper[4818]: I1203 07:16:50.738029 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:16:50 crc kubenswrapper[4818]: E1203 07:16:50.738997 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:17:04 crc kubenswrapper[4818]: I1203 07:17:04.738654 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:17:04 crc kubenswrapper[4818]: E1203 07:17:04.740057 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:17:16 crc kubenswrapper[4818]: I1203 07:17:16.738482 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:17:16 crc kubenswrapper[4818]: E1203 07:17:16.739370 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:17:29 crc kubenswrapper[4818]: I1203 07:17:29.738478 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:17:29 crc kubenswrapper[4818]: E1203 07:17:29.739847 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:17:43 crc kubenswrapper[4818]: I1203 07:17:43.737597 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:17:43 crc kubenswrapper[4818]: E1203 07:17:43.739302 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:17:55 crc kubenswrapper[4818]: I1203 07:17:55.738126 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:17:55 crc kubenswrapper[4818]: E1203 07:17:55.739089 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:18:06 crc kubenswrapper[4818]: I1203 07:18:06.738782 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:18:06 crc kubenswrapper[4818]: E1203 07:18:06.739724 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:18:18 crc kubenswrapper[4818]: I1203 07:18:18.738718 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:18:18 crc kubenswrapper[4818]: E1203 07:18:18.740932 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:18:32 crc kubenswrapper[4818]: I1203 07:18:32.755893 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:18:32 crc kubenswrapper[4818]: E1203 07:18:32.757207 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:18:45 crc kubenswrapper[4818]: I1203 07:18:45.737344 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:18:45 crc kubenswrapper[4818]: E1203 07:18:45.738282 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:19:00 crc kubenswrapper[4818]: I1203 07:19:00.738366 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:19:00 crc kubenswrapper[4818]: E1203 07:19:00.739735 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:19:12 crc kubenswrapper[4818]: I1203 07:19:12.745751 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:19:12 crc kubenswrapper[4818]: E1203 07:19:12.747279 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:19:24 crc kubenswrapper[4818]: I1203 07:19:24.737363 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:19:24 crc kubenswrapper[4818]: E1203 07:19:24.738212 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:19:38 crc kubenswrapper[4818]: I1203 07:19:38.738605 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:19:38 crc kubenswrapper[4818]: E1203 07:19:38.739641 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:19:49 crc kubenswrapper[4818]: I1203 07:19:49.757266 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:19:49 crc kubenswrapper[4818]: E1203 07:19:49.757921 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.041408 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wttq5"] Dec 03 07:19:53 crc kubenswrapper[4818]: E1203 07:19:53.042732 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32413695-c405-402e-bc2b-1eda396326f7" containerName="registry-server" Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.042764 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="32413695-c405-402e-bc2b-1eda396326f7" containerName="registry-server" Dec 03 07:19:53 crc kubenswrapper[4818]: E1203 07:19:53.042823 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32413695-c405-402e-bc2b-1eda396326f7" containerName="extract-utilities" Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.042859 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="32413695-c405-402e-bc2b-1eda396326f7" containerName="extract-utilities" Dec 03 07:19:53 crc kubenswrapper[4818]: E1203 07:19:53.042884 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32413695-c405-402e-bc2b-1eda396326f7" containerName="extract-content" Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.042898 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="32413695-c405-402e-bc2b-1eda396326f7" containerName="extract-content" Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.043283 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="32413695-c405-402e-bc2b-1eda396326f7" containerName="registry-server" Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.045919 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wttq5" Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.052017 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wttq5"] Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.121841 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/746ddfec-5474-4eec-a212-645aedd2f708-utilities\") pod \"redhat-marketplace-wttq5\" (UID: \"746ddfec-5474-4eec-a212-645aedd2f708\") " pod="openshift-marketplace/redhat-marketplace-wttq5" Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.121971 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/746ddfec-5474-4eec-a212-645aedd2f708-catalog-content\") pod \"redhat-marketplace-wttq5\" (UID: \"746ddfec-5474-4eec-a212-645aedd2f708\") " pod="openshift-marketplace/redhat-marketplace-wttq5" Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.122236 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jg4t\" (UniqueName: \"kubernetes.io/projected/746ddfec-5474-4eec-a212-645aedd2f708-kube-api-access-2jg4t\") pod \"redhat-marketplace-wttq5\" (UID: \"746ddfec-5474-4eec-a212-645aedd2f708\") " pod="openshift-marketplace/redhat-marketplace-wttq5" Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.224290 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/746ddfec-5474-4eec-a212-645aedd2f708-utilities\") pod \"redhat-marketplace-wttq5\" (UID: \"746ddfec-5474-4eec-a212-645aedd2f708\") " pod="openshift-marketplace/redhat-marketplace-wttq5" Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.224675 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/746ddfec-5474-4eec-a212-645aedd2f708-catalog-content\") pod \"redhat-marketplace-wttq5\" (UID: \"746ddfec-5474-4eec-a212-645aedd2f708\") " pod="openshift-marketplace/redhat-marketplace-wttq5" Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.224865 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jg4t\" (UniqueName: \"kubernetes.io/projected/746ddfec-5474-4eec-a212-645aedd2f708-kube-api-access-2jg4t\") pod \"redhat-marketplace-wttq5\" (UID: \"746ddfec-5474-4eec-a212-645aedd2f708\") " pod="openshift-marketplace/redhat-marketplace-wttq5" Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.225769 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/746ddfec-5474-4eec-a212-645aedd2f708-utilities\") pod \"redhat-marketplace-wttq5\" (UID: \"746ddfec-5474-4eec-a212-645aedd2f708\") " pod="openshift-marketplace/redhat-marketplace-wttq5" Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.226172 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/746ddfec-5474-4eec-a212-645aedd2f708-catalog-content\") pod \"redhat-marketplace-wttq5\" (UID: \"746ddfec-5474-4eec-a212-645aedd2f708\") " pod="openshift-marketplace/redhat-marketplace-wttq5" Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.245096 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jg4t\" (UniqueName: \"kubernetes.io/projected/746ddfec-5474-4eec-a212-645aedd2f708-kube-api-access-2jg4t\") pod \"redhat-marketplace-wttq5\" (UID: \"746ddfec-5474-4eec-a212-645aedd2f708\") " pod="openshift-marketplace/redhat-marketplace-wttq5" Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.385120 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wttq5" Dec 03 07:19:53 crc kubenswrapper[4818]: I1203 07:19:53.908866 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wttq5"] Dec 03 07:19:54 crc kubenswrapper[4818]: I1203 07:19:54.648409 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wttq5" event={"ID":"746ddfec-5474-4eec-a212-645aedd2f708","Type":"ContainerStarted","Data":"cf0f74f30bc9e16b82cdb6c7c6472282e9bacb12d284dde725a82d0681ec897e"} Dec 03 07:19:56 crc kubenswrapper[4818]: I1203 07:19:56.671936 4818 generic.go:334] "Generic (PLEG): container finished" podID="746ddfec-5474-4eec-a212-645aedd2f708" containerID="9fdff7690ff7c4ba14d1fcdc643c8cb0d9b0deeef67fd4114984efb69c15cf2f" exitCode=0 Dec 03 07:19:56 crc kubenswrapper[4818]: I1203 07:19:56.672009 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wttq5" event={"ID":"746ddfec-5474-4eec-a212-645aedd2f708","Type":"ContainerDied","Data":"9fdff7690ff7c4ba14d1fcdc643c8cb0d9b0deeef67fd4114984efb69c15cf2f"} Dec 03 07:19:57 crc kubenswrapper[4818]: I1203 07:19:57.690326 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 07:19:59 crc kubenswrapper[4818]: I1203 07:19:59.711912 4818 generic.go:334] "Generic (PLEG): container finished" podID="746ddfec-5474-4eec-a212-645aedd2f708" containerID="f67d71d9b2455c7d07dea1d8fd11fa3b8b0f263c0157642ee9206ea4a3874ca2" exitCode=0 Dec 03 07:19:59 crc kubenswrapper[4818]: I1203 07:19:59.712030 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wttq5" event={"ID":"746ddfec-5474-4eec-a212-645aedd2f708","Type":"ContainerDied","Data":"f67d71d9b2455c7d07dea1d8fd11fa3b8b0f263c0157642ee9206ea4a3874ca2"} Dec 03 07:20:00 crc kubenswrapper[4818]: I1203 07:20:00.723693 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wttq5" event={"ID":"746ddfec-5474-4eec-a212-645aedd2f708","Type":"ContainerStarted","Data":"53bec6b03a0f12d01ab3d775bcf3b01f6dbb5d9a5a17dfdea2b0c97b0139f3b1"} Dec 03 07:20:00 crc kubenswrapper[4818]: I1203 07:20:00.749624 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wttq5" podStartSLOduration=5.291912936 podStartE2EDuration="7.749574804s" podCreationTimestamp="2025-12-03 07:19:53 +0000 UTC" firstStartedPulling="2025-12-03 07:19:57.690023098 +0000 UTC m=+3155.381631850" lastFinishedPulling="2025-12-03 07:20:00.147684966 +0000 UTC m=+3157.839293718" observedRunningTime="2025-12-03 07:20:00.742094617 +0000 UTC m=+3158.433703379" watchObservedRunningTime="2025-12-03 07:20:00.749574804 +0000 UTC m=+3158.441183556" Dec 03 07:20:02 crc kubenswrapper[4818]: I1203 07:20:02.745247 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:20:02 crc kubenswrapper[4818]: E1203 07:20:02.745921 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:20:03 crc kubenswrapper[4818]: I1203 07:20:03.386556 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wttq5" Dec 03 07:20:03 crc kubenswrapper[4818]: I1203 07:20:03.386612 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wttq5" Dec 03 07:20:03 crc kubenswrapper[4818]: I1203 07:20:03.455769 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wttq5" Dec 03 07:20:13 crc kubenswrapper[4818]: I1203 07:20:13.464927 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wttq5" Dec 03 07:20:13 crc kubenswrapper[4818]: I1203 07:20:13.538326 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wttq5"] Dec 03 07:20:13 crc kubenswrapper[4818]: I1203 07:20:13.864712 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wttq5" podUID="746ddfec-5474-4eec-a212-645aedd2f708" containerName="registry-server" containerID="cri-o://53bec6b03a0f12d01ab3d775bcf3b01f6dbb5d9a5a17dfdea2b0c97b0139f3b1" gracePeriod=2 Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.373807 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wttq5" Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.408765 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/746ddfec-5474-4eec-a212-645aedd2f708-catalog-content\") pod \"746ddfec-5474-4eec-a212-645aedd2f708\" (UID: \"746ddfec-5474-4eec-a212-645aedd2f708\") " Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.408831 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jg4t\" (UniqueName: \"kubernetes.io/projected/746ddfec-5474-4eec-a212-645aedd2f708-kube-api-access-2jg4t\") pod \"746ddfec-5474-4eec-a212-645aedd2f708\" (UID: \"746ddfec-5474-4eec-a212-645aedd2f708\") " Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.408862 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/746ddfec-5474-4eec-a212-645aedd2f708-utilities\") pod \"746ddfec-5474-4eec-a212-645aedd2f708\" (UID: \"746ddfec-5474-4eec-a212-645aedd2f708\") " Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.415701 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/746ddfec-5474-4eec-a212-645aedd2f708-utilities" (OuterVolumeSpecName: "utilities") pod "746ddfec-5474-4eec-a212-645aedd2f708" (UID: "746ddfec-5474-4eec-a212-645aedd2f708"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.418121 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/746ddfec-5474-4eec-a212-645aedd2f708-kube-api-access-2jg4t" (OuterVolumeSpecName: "kube-api-access-2jg4t") pod "746ddfec-5474-4eec-a212-645aedd2f708" (UID: "746ddfec-5474-4eec-a212-645aedd2f708"). InnerVolumeSpecName "kube-api-access-2jg4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.431522 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/746ddfec-5474-4eec-a212-645aedd2f708-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "746ddfec-5474-4eec-a212-645aedd2f708" (UID: "746ddfec-5474-4eec-a212-645aedd2f708"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.516009 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/746ddfec-5474-4eec-a212-645aedd2f708-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.516047 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jg4t\" (UniqueName: \"kubernetes.io/projected/746ddfec-5474-4eec-a212-645aedd2f708-kube-api-access-2jg4t\") on node \"crc\" DevicePath \"\"" Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.516058 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/746ddfec-5474-4eec-a212-645aedd2f708-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.878513 4818 generic.go:334] "Generic (PLEG): container finished" podID="746ddfec-5474-4eec-a212-645aedd2f708" containerID="53bec6b03a0f12d01ab3d775bcf3b01f6dbb5d9a5a17dfdea2b0c97b0139f3b1" exitCode=0 Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.878579 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wttq5" event={"ID":"746ddfec-5474-4eec-a212-645aedd2f708","Type":"ContainerDied","Data":"53bec6b03a0f12d01ab3d775bcf3b01f6dbb5d9a5a17dfdea2b0c97b0139f3b1"} Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.878622 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wttq5" Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.878648 4818 scope.go:117] "RemoveContainer" containerID="53bec6b03a0f12d01ab3d775bcf3b01f6dbb5d9a5a17dfdea2b0c97b0139f3b1" Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.878629 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wttq5" event={"ID":"746ddfec-5474-4eec-a212-645aedd2f708","Type":"ContainerDied","Data":"cf0f74f30bc9e16b82cdb6c7c6472282e9bacb12d284dde725a82d0681ec897e"} Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.910661 4818 scope.go:117] "RemoveContainer" containerID="f67d71d9b2455c7d07dea1d8fd11fa3b8b0f263c0157642ee9206ea4a3874ca2" Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.917073 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wttq5"] Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.926751 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wttq5"] Dec 03 07:20:14 crc kubenswrapper[4818]: I1203 07:20:14.942547 4818 scope.go:117] "RemoveContainer" containerID="9fdff7690ff7c4ba14d1fcdc643c8cb0d9b0deeef67fd4114984efb69c15cf2f" Dec 03 07:20:15 crc kubenswrapper[4818]: I1203 07:20:15.009911 4818 scope.go:117] "RemoveContainer" containerID="53bec6b03a0f12d01ab3d775bcf3b01f6dbb5d9a5a17dfdea2b0c97b0139f3b1" Dec 03 07:20:15 crc kubenswrapper[4818]: E1203 07:20:15.010440 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53bec6b03a0f12d01ab3d775bcf3b01f6dbb5d9a5a17dfdea2b0c97b0139f3b1\": container with ID starting with 53bec6b03a0f12d01ab3d775bcf3b01f6dbb5d9a5a17dfdea2b0c97b0139f3b1 not found: ID does not exist" containerID="53bec6b03a0f12d01ab3d775bcf3b01f6dbb5d9a5a17dfdea2b0c97b0139f3b1" Dec 03 07:20:15 crc kubenswrapper[4818]: I1203 07:20:15.010481 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53bec6b03a0f12d01ab3d775bcf3b01f6dbb5d9a5a17dfdea2b0c97b0139f3b1"} err="failed to get container status \"53bec6b03a0f12d01ab3d775bcf3b01f6dbb5d9a5a17dfdea2b0c97b0139f3b1\": rpc error: code = NotFound desc = could not find container \"53bec6b03a0f12d01ab3d775bcf3b01f6dbb5d9a5a17dfdea2b0c97b0139f3b1\": container with ID starting with 53bec6b03a0f12d01ab3d775bcf3b01f6dbb5d9a5a17dfdea2b0c97b0139f3b1 not found: ID does not exist" Dec 03 07:20:15 crc kubenswrapper[4818]: I1203 07:20:15.010506 4818 scope.go:117] "RemoveContainer" containerID="f67d71d9b2455c7d07dea1d8fd11fa3b8b0f263c0157642ee9206ea4a3874ca2" Dec 03 07:20:15 crc kubenswrapper[4818]: E1203 07:20:15.010786 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f67d71d9b2455c7d07dea1d8fd11fa3b8b0f263c0157642ee9206ea4a3874ca2\": container with ID starting with f67d71d9b2455c7d07dea1d8fd11fa3b8b0f263c0157642ee9206ea4a3874ca2 not found: ID does not exist" containerID="f67d71d9b2455c7d07dea1d8fd11fa3b8b0f263c0157642ee9206ea4a3874ca2" Dec 03 07:20:15 crc kubenswrapper[4818]: I1203 07:20:15.010836 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f67d71d9b2455c7d07dea1d8fd11fa3b8b0f263c0157642ee9206ea4a3874ca2"} err="failed to get container status \"f67d71d9b2455c7d07dea1d8fd11fa3b8b0f263c0157642ee9206ea4a3874ca2\": rpc error: code = NotFound desc = could not find container \"f67d71d9b2455c7d07dea1d8fd11fa3b8b0f263c0157642ee9206ea4a3874ca2\": container with ID starting with f67d71d9b2455c7d07dea1d8fd11fa3b8b0f263c0157642ee9206ea4a3874ca2 not found: ID does not exist" Dec 03 07:20:15 crc kubenswrapper[4818]: I1203 07:20:15.010862 4818 scope.go:117] "RemoveContainer" containerID="9fdff7690ff7c4ba14d1fcdc643c8cb0d9b0deeef67fd4114984efb69c15cf2f" Dec 03 07:20:15 crc kubenswrapper[4818]: E1203 07:20:15.011329 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fdff7690ff7c4ba14d1fcdc643c8cb0d9b0deeef67fd4114984efb69c15cf2f\": container with ID starting with 9fdff7690ff7c4ba14d1fcdc643c8cb0d9b0deeef67fd4114984efb69c15cf2f not found: ID does not exist" containerID="9fdff7690ff7c4ba14d1fcdc643c8cb0d9b0deeef67fd4114984efb69c15cf2f" Dec 03 07:20:15 crc kubenswrapper[4818]: I1203 07:20:15.011369 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fdff7690ff7c4ba14d1fcdc643c8cb0d9b0deeef67fd4114984efb69c15cf2f"} err="failed to get container status \"9fdff7690ff7c4ba14d1fcdc643c8cb0d9b0deeef67fd4114984efb69c15cf2f\": rpc error: code = NotFound desc = could not find container \"9fdff7690ff7c4ba14d1fcdc643c8cb0d9b0deeef67fd4114984efb69c15cf2f\": container with ID starting with 9fdff7690ff7c4ba14d1fcdc643c8cb0d9b0deeef67fd4114984efb69c15cf2f not found: ID does not exist" Dec 03 07:20:16 crc kubenswrapper[4818]: I1203 07:20:16.737763 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:20:16 crc kubenswrapper[4818]: E1203 07:20:16.738549 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:20:16 crc kubenswrapper[4818]: I1203 07:20:16.755872 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="746ddfec-5474-4eec-a212-645aedd2f708" path="/var/lib/kubelet/pods/746ddfec-5474-4eec-a212-645aedd2f708/volumes" Dec 03 07:20:29 crc kubenswrapper[4818]: I1203 07:20:29.738122 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:20:29 crc kubenswrapper[4818]: E1203 07:20:29.738921 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:20:41 crc kubenswrapper[4818]: I1203 07:20:41.737532 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:20:41 crc kubenswrapper[4818]: E1203 07:20:41.738174 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:20:56 crc kubenswrapper[4818]: I1203 07:20:56.739525 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:20:56 crc kubenswrapper[4818]: E1203 07:20:56.740277 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:21:10 crc kubenswrapper[4818]: I1203 07:21:10.738746 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:21:10 crc kubenswrapper[4818]: E1203 07:21:10.739773 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:21:23 crc kubenswrapper[4818]: I1203 07:21:23.738160 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:21:24 crc kubenswrapper[4818]: I1203 07:21:24.643569 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"fe2c3fb698cd81e0da40cc4153b5a00fd512ba89b40ba70448f8ea1e0ace6325"} Dec 03 07:23:43 crc kubenswrapper[4818]: I1203 07:23:43.302220 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:23:43 crc kubenswrapper[4818]: I1203 07:23:43.303721 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:24:13 crc kubenswrapper[4818]: I1203 07:24:13.302359 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:24:13 crc kubenswrapper[4818]: I1203 07:24:13.302901 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:24:43 crc kubenswrapper[4818]: I1203 07:24:43.302461 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:24:43 crc kubenswrapper[4818]: I1203 07:24:43.303102 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:24:43 crc kubenswrapper[4818]: I1203 07:24:43.303157 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 07:24:43 crc kubenswrapper[4818]: I1203 07:24:43.303983 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fe2c3fb698cd81e0da40cc4153b5a00fd512ba89b40ba70448f8ea1e0ace6325"} pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:24:43 crc kubenswrapper[4818]: I1203 07:24:43.304609 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" containerID="cri-o://fe2c3fb698cd81e0da40cc4153b5a00fd512ba89b40ba70448f8ea1e0ace6325" gracePeriod=600 Dec 03 07:24:43 crc kubenswrapper[4818]: I1203 07:24:43.690302 4818 generic.go:334] "Generic (PLEG): container finished" podID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerID="fe2c3fb698cd81e0da40cc4153b5a00fd512ba89b40ba70448f8ea1e0ace6325" exitCode=0 Dec 03 07:24:43 crc kubenswrapper[4818]: I1203 07:24:43.690361 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerDied","Data":"fe2c3fb698cd81e0da40cc4153b5a00fd512ba89b40ba70448f8ea1e0ace6325"} Dec 03 07:24:43 crc kubenswrapper[4818]: I1203 07:24:43.690684 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020"} Dec 03 07:24:43 crc kubenswrapper[4818]: I1203 07:24:43.690736 4818 scope.go:117] "RemoveContainer" containerID="9a5459c0bec10022abcf4012f64a6eb3de782e53e35cebb77eac50ee95afdda0" Dec 03 07:25:08 crc kubenswrapper[4818]: I1203 07:25:08.513441 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fv8rv"] Dec 03 07:25:08 crc kubenswrapper[4818]: E1203 07:25:08.514602 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="746ddfec-5474-4eec-a212-645aedd2f708" containerName="extract-utilities" Dec 03 07:25:08 crc kubenswrapper[4818]: I1203 07:25:08.514620 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="746ddfec-5474-4eec-a212-645aedd2f708" containerName="extract-utilities" Dec 03 07:25:08 crc kubenswrapper[4818]: E1203 07:25:08.514650 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="746ddfec-5474-4eec-a212-645aedd2f708" containerName="registry-server" Dec 03 07:25:08 crc kubenswrapper[4818]: I1203 07:25:08.514668 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="746ddfec-5474-4eec-a212-645aedd2f708" containerName="registry-server" Dec 03 07:25:08 crc kubenswrapper[4818]: E1203 07:25:08.514685 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="746ddfec-5474-4eec-a212-645aedd2f708" containerName="extract-content" Dec 03 07:25:08 crc kubenswrapper[4818]: I1203 07:25:08.514693 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="746ddfec-5474-4eec-a212-645aedd2f708" containerName="extract-content" Dec 03 07:25:08 crc kubenswrapper[4818]: I1203 07:25:08.514985 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="746ddfec-5474-4eec-a212-645aedd2f708" containerName="registry-server" Dec 03 07:25:08 crc kubenswrapper[4818]: I1203 07:25:08.516669 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fv8rv" Dec 03 07:25:08 crc kubenswrapper[4818]: I1203 07:25:08.525669 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fv8rv"] Dec 03 07:25:08 crc kubenswrapper[4818]: I1203 07:25:08.676912 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5052d2-0875-421c-8f43-e979da7f8681-utilities\") pod \"certified-operators-fv8rv\" (UID: \"ce5052d2-0875-421c-8f43-e979da7f8681\") " pod="openshift-marketplace/certified-operators-fv8rv" Dec 03 07:25:08 crc kubenswrapper[4818]: I1203 07:25:08.677186 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5052d2-0875-421c-8f43-e979da7f8681-catalog-content\") pod \"certified-operators-fv8rv\" (UID: \"ce5052d2-0875-421c-8f43-e979da7f8681\") " pod="openshift-marketplace/certified-operators-fv8rv" Dec 03 07:25:08 crc kubenswrapper[4818]: I1203 07:25:08.677360 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgdjw\" (UniqueName: \"kubernetes.io/projected/ce5052d2-0875-421c-8f43-e979da7f8681-kube-api-access-vgdjw\") pod \"certified-operators-fv8rv\" (UID: \"ce5052d2-0875-421c-8f43-e979da7f8681\") " pod="openshift-marketplace/certified-operators-fv8rv" Dec 03 07:25:08 crc kubenswrapper[4818]: I1203 07:25:08.779398 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5052d2-0875-421c-8f43-e979da7f8681-utilities\") pod \"certified-operators-fv8rv\" (UID: \"ce5052d2-0875-421c-8f43-e979da7f8681\") " pod="openshift-marketplace/certified-operators-fv8rv" Dec 03 07:25:08 crc kubenswrapper[4818]: I1203 07:25:08.779486 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5052d2-0875-421c-8f43-e979da7f8681-catalog-content\") pod \"certified-operators-fv8rv\" (UID: \"ce5052d2-0875-421c-8f43-e979da7f8681\") " pod="openshift-marketplace/certified-operators-fv8rv" Dec 03 07:25:08 crc kubenswrapper[4818]: I1203 07:25:08.779578 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgdjw\" (UniqueName: \"kubernetes.io/projected/ce5052d2-0875-421c-8f43-e979da7f8681-kube-api-access-vgdjw\") pod \"certified-operators-fv8rv\" (UID: \"ce5052d2-0875-421c-8f43-e979da7f8681\") " pod="openshift-marketplace/certified-operators-fv8rv" Dec 03 07:25:08 crc kubenswrapper[4818]: I1203 07:25:08.780323 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5052d2-0875-421c-8f43-e979da7f8681-catalog-content\") pod \"certified-operators-fv8rv\" (UID: \"ce5052d2-0875-421c-8f43-e979da7f8681\") " pod="openshift-marketplace/certified-operators-fv8rv" Dec 03 07:25:08 crc kubenswrapper[4818]: I1203 07:25:08.780333 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5052d2-0875-421c-8f43-e979da7f8681-utilities\") pod \"certified-operators-fv8rv\" (UID: \"ce5052d2-0875-421c-8f43-e979da7f8681\") " pod="openshift-marketplace/certified-operators-fv8rv" Dec 03 07:25:08 crc kubenswrapper[4818]: I1203 07:25:08.801964 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgdjw\" (UniqueName: \"kubernetes.io/projected/ce5052d2-0875-421c-8f43-e979da7f8681-kube-api-access-vgdjw\") pod \"certified-operators-fv8rv\" (UID: \"ce5052d2-0875-421c-8f43-e979da7f8681\") " pod="openshift-marketplace/certified-operators-fv8rv" Dec 03 07:25:08 crc kubenswrapper[4818]: I1203 07:25:08.838236 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fv8rv" Dec 03 07:25:09 crc kubenswrapper[4818]: I1203 07:25:09.424998 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fv8rv"] Dec 03 07:25:10 crc kubenswrapper[4818]: I1203 07:25:10.053382 4818 generic.go:334] "Generic (PLEG): container finished" podID="ce5052d2-0875-421c-8f43-e979da7f8681" containerID="3a78c83f5019e88325de3c4eebe529076210193c6ed781097a5586c0fff93fa5" exitCode=0 Dec 03 07:25:10 crc kubenswrapper[4818]: I1203 07:25:10.053451 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv8rv" event={"ID":"ce5052d2-0875-421c-8f43-e979da7f8681","Type":"ContainerDied","Data":"3a78c83f5019e88325de3c4eebe529076210193c6ed781097a5586c0fff93fa5"} Dec 03 07:25:10 crc kubenswrapper[4818]: I1203 07:25:10.053496 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv8rv" event={"ID":"ce5052d2-0875-421c-8f43-e979da7f8681","Type":"ContainerStarted","Data":"d9790df6aa69f96b2a71b26ffa81e2c045e4f115ac17013b1731d640972a26c1"} Dec 03 07:25:10 crc kubenswrapper[4818]: I1203 07:25:10.056372 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 07:25:11 crc kubenswrapper[4818]: I1203 07:25:11.067734 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv8rv" event={"ID":"ce5052d2-0875-421c-8f43-e979da7f8681","Type":"ContainerStarted","Data":"1d25c4c4a4e2d5949868eb08e3bd7f32abaede97557ea2775664da1838d5c3e7"} Dec 03 07:25:12 crc kubenswrapper[4818]: I1203 07:25:12.079613 4818 generic.go:334] "Generic (PLEG): container finished" podID="ce5052d2-0875-421c-8f43-e979da7f8681" containerID="1d25c4c4a4e2d5949868eb08e3bd7f32abaede97557ea2775664da1838d5c3e7" exitCode=0 Dec 03 07:25:12 crc kubenswrapper[4818]: I1203 07:25:12.079662 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv8rv" event={"ID":"ce5052d2-0875-421c-8f43-e979da7f8681","Type":"ContainerDied","Data":"1d25c4c4a4e2d5949868eb08e3bd7f32abaede97557ea2775664da1838d5c3e7"} Dec 03 07:25:12 crc kubenswrapper[4818]: I1203 07:25:12.985302 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lj86j"] Dec 03 07:25:12 crc kubenswrapper[4818]: I1203 07:25:12.987499 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lj86j" Dec 03 07:25:13 crc kubenswrapper[4818]: I1203 07:25:13.010503 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lj86j"] Dec 03 07:25:13 crc kubenswrapper[4818]: I1203 07:25:13.088713 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv8rv" event={"ID":"ce5052d2-0875-421c-8f43-e979da7f8681","Type":"ContainerStarted","Data":"dd43fcd4bf76c596ee5566428d60eed3ce0171c970170ce7e51bd0574a87c8f1"} Dec 03 07:25:13 crc kubenswrapper[4818]: I1203 07:25:13.106182 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z62gn\" (UniqueName: \"kubernetes.io/projected/8d18581c-b5a3-4132-afcc-71bced63091b-kube-api-access-z62gn\") pod \"redhat-operators-lj86j\" (UID: \"8d18581c-b5a3-4132-afcc-71bced63091b\") " pod="openshift-marketplace/redhat-operators-lj86j" Dec 03 07:25:13 crc kubenswrapper[4818]: I1203 07:25:13.106237 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d18581c-b5a3-4132-afcc-71bced63091b-catalog-content\") pod \"redhat-operators-lj86j\" (UID: \"8d18581c-b5a3-4132-afcc-71bced63091b\") " pod="openshift-marketplace/redhat-operators-lj86j" Dec 03 07:25:13 crc kubenswrapper[4818]: I1203 07:25:13.106618 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d18581c-b5a3-4132-afcc-71bced63091b-utilities\") pod \"redhat-operators-lj86j\" (UID: \"8d18581c-b5a3-4132-afcc-71bced63091b\") " pod="openshift-marketplace/redhat-operators-lj86j" Dec 03 07:25:13 crc kubenswrapper[4818]: I1203 07:25:13.110944 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fv8rv" podStartSLOduration=2.522777567 podStartE2EDuration="5.110916402s" podCreationTimestamp="2025-12-03 07:25:08 +0000 UTC" firstStartedPulling="2025-12-03 07:25:10.056037863 +0000 UTC m=+3467.747646615" lastFinishedPulling="2025-12-03 07:25:12.644176678 +0000 UTC m=+3470.335785450" observedRunningTime="2025-12-03 07:25:13.104072342 +0000 UTC m=+3470.795681084" watchObservedRunningTime="2025-12-03 07:25:13.110916402 +0000 UTC m=+3470.802525154" Dec 03 07:25:13 crc kubenswrapper[4818]: I1203 07:25:13.208950 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d18581c-b5a3-4132-afcc-71bced63091b-utilities\") pod \"redhat-operators-lj86j\" (UID: \"8d18581c-b5a3-4132-afcc-71bced63091b\") " pod="openshift-marketplace/redhat-operators-lj86j" Dec 03 07:25:13 crc kubenswrapper[4818]: I1203 07:25:13.209053 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z62gn\" (UniqueName: \"kubernetes.io/projected/8d18581c-b5a3-4132-afcc-71bced63091b-kube-api-access-z62gn\") pod \"redhat-operators-lj86j\" (UID: \"8d18581c-b5a3-4132-afcc-71bced63091b\") " pod="openshift-marketplace/redhat-operators-lj86j" Dec 03 07:25:13 crc kubenswrapper[4818]: I1203 07:25:13.209077 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d18581c-b5a3-4132-afcc-71bced63091b-catalog-content\") pod \"redhat-operators-lj86j\" (UID: \"8d18581c-b5a3-4132-afcc-71bced63091b\") " pod="openshift-marketplace/redhat-operators-lj86j" Dec 03 07:25:13 crc kubenswrapper[4818]: I1203 07:25:13.209673 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d18581c-b5a3-4132-afcc-71bced63091b-catalog-content\") pod \"redhat-operators-lj86j\" (UID: \"8d18581c-b5a3-4132-afcc-71bced63091b\") " pod="openshift-marketplace/redhat-operators-lj86j" Dec 03 07:25:13 crc kubenswrapper[4818]: I1203 07:25:13.210236 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d18581c-b5a3-4132-afcc-71bced63091b-utilities\") pod \"redhat-operators-lj86j\" (UID: \"8d18581c-b5a3-4132-afcc-71bced63091b\") " pod="openshift-marketplace/redhat-operators-lj86j" Dec 03 07:25:13 crc kubenswrapper[4818]: I1203 07:25:13.234943 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z62gn\" (UniqueName: \"kubernetes.io/projected/8d18581c-b5a3-4132-afcc-71bced63091b-kube-api-access-z62gn\") pod \"redhat-operators-lj86j\" (UID: \"8d18581c-b5a3-4132-afcc-71bced63091b\") " pod="openshift-marketplace/redhat-operators-lj86j" Dec 03 07:25:13 crc kubenswrapper[4818]: I1203 07:25:13.340781 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lj86j" Dec 03 07:25:13 crc kubenswrapper[4818]: I1203 07:25:13.832465 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lj86j"] Dec 03 07:25:14 crc kubenswrapper[4818]: I1203 07:25:14.099065 4818 generic.go:334] "Generic (PLEG): container finished" podID="8d18581c-b5a3-4132-afcc-71bced63091b" containerID="0d465c3eac67ddd1fe0b084045b20a8af4746b13cf450d24240881bd36cdf74a" exitCode=0 Dec 03 07:25:14 crc kubenswrapper[4818]: I1203 07:25:14.099120 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lj86j" event={"ID":"8d18581c-b5a3-4132-afcc-71bced63091b","Type":"ContainerDied","Data":"0d465c3eac67ddd1fe0b084045b20a8af4746b13cf450d24240881bd36cdf74a"} Dec 03 07:25:14 crc kubenswrapper[4818]: I1203 07:25:14.100510 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lj86j" event={"ID":"8d18581c-b5a3-4132-afcc-71bced63091b","Type":"ContainerStarted","Data":"d7ae9c2cb2367c086f4ef3a2299b8ed379b86254cd2b52d529ca067a42af7646"} Dec 03 07:25:15 crc kubenswrapper[4818]: I1203 07:25:15.110791 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lj86j" event={"ID":"8d18581c-b5a3-4132-afcc-71bced63091b","Type":"ContainerStarted","Data":"b90558580c10f4ba9df13dd489b076af606c0623ef2e525fe0f69d711f25de3b"} Dec 03 07:25:18 crc kubenswrapper[4818]: I1203 07:25:18.138191 4818 generic.go:334] "Generic (PLEG): container finished" podID="8d18581c-b5a3-4132-afcc-71bced63091b" containerID="b90558580c10f4ba9df13dd489b076af606c0623ef2e525fe0f69d711f25de3b" exitCode=0 Dec 03 07:25:18 crc kubenswrapper[4818]: I1203 07:25:18.138283 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lj86j" event={"ID":"8d18581c-b5a3-4132-afcc-71bced63091b","Type":"ContainerDied","Data":"b90558580c10f4ba9df13dd489b076af606c0623ef2e525fe0f69d711f25de3b"} Dec 03 07:25:18 crc kubenswrapper[4818]: I1203 07:25:18.839252 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fv8rv" Dec 03 07:25:18 crc kubenswrapper[4818]: I1203 07:25:18.852944 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fv8rv" Dec 03 07:25:19 crc kubenswrapper[4818]: I1203 07:25:19.080872 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fv8rv" Dec 03 07:25:19 crc kubenswrapper[4818]: I1203 07:25:19.201653 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fv8rv" Dec 03 07:25:20 crc kubenswrapper[4818]: I1203 07:25:20.158469 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lj86j" event={"ID":"8d18581c-b5a3-4132-afcc-71bced63091b","Type":"ContainerStarted","Data":"e9f0ec583ce4c81e4e42d47520b0d800a19a03bf38c8c27c4ad385abd7203cf9"} Dec 03 07:25:20 crc kubenswrapper[4818]: I1203 07:25:20.192067 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lj86j" podStartSLOduration=3.112946122 podStartE2EDuration="8.192026702s" podCreationTimestamp="2025-12-03 07:25:12 +0000 UTC" firstStartedPulling="2025-12-03 07:25:14.100807675 +0000 UTC m=+3471.792416427" lastFinishedPulling="2025-12-03 07:25:19.179888255 +0000 UTC m=+3476.871497007" observedRunningTime="2025-12-03 07:25:20.187369476 +0000 UTC m=+3477.878978248" watchObservedRunningTime="2025-12-03 07:25:20.192026702 +0000 UTC m=+3477.883635474" Dec 03 07:25:20 crc kubenswrapper[4818]: I1203 07:25:20.506300 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fv8rv"] Dec 03 07:25:22 crc kubenswrapper[4818]: I1203 07:25:22.176742 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fv8rv" podUID="ce5052d2-0875-421c-8f43-e979da7f8681" containerName="registry-server" containerID="cri-o://dd43fcd4bf76c596ee5566428d60eed3ce0171c970170ce7e51bd0574a87c8f1" gracePeriod=2 Dec 03 07:25:22 crc kubenswrapper[4818]: E1203 07:25:22.424752 4818 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce5052d2_0875_421c_8f43_e979da7f8681.slice/crio-dd43fcd4bf76c596ee5566428d60eed3ce0171c970170ce7e51bd0574a87c8f1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce5052d2_0875_421c_8f43_e979da7f8681.slice/crio-conmon-dd43fcd4bf76c596ee5566428d60eed3ce0171c970170ce7e51bd0574a87c8f1.scope\": RecentStats: unable to find data in memory cache]" Dec 03 07:25:22 crc kubenswrapper[4818]: I1203 07:25:22.791572 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fv8rv" Dec 03 07:25:22 crc kubenswrapper[4818]: I1203 07:25:22.911996 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgdjw\" (UniqueName: \"kubernetes.io/projected/ce5052d2-0875-421c-8f43-e979da7f8681-kube-api-access-vgdjw\") pod \"ce5052d2-0875-421c-8f43-e979da7f8681\" (UID: \"ce5052d2-0875-421c-8f43-e979da7f8681\") " Dec 03 07:25:22 crc kubenswrapper[4818]: I1203 07:25:22.912072 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5052d2-0875-421c-8f43-e979da7f8681-utilities\") pod \"ce5052d2-0875-421c-8f43-e979da7f8681\" (UID: \"ce5052d2-0875-421c-8f43-e979da7f8681\") " Dec 03 07:25:22 crc kubenswrapper[4818]: I1203 07:25:22.912117 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5052d2-0875-421c-8f43-e979da7f8681-catalog-content\") pod \"ce5052d2-0875-421c-8f43-e979da7f8681\" (UID: \"ce5052d2-0875-421c-8f43-e979da7f8681\") " Dec 03 07:25:22 crc kubenswrapper[4818]: I1203 07:25:22.912625 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce5052d2-0875-421c-8f43-e979da7f8681-utilities" (OuterVolumeSpecName: "utilities") pod "ce5052d2-0875-421c-8f43-e979da7f8681" (UID: "ce5052d2-0875-421c-8f43-e979da7f8681"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:25:22 crc kubenswrapper[4818]: I1203 07:25:22.926390 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce5052d2-0875-421c-8f43-e979da7f8681-kube-api-access-vgdjw" (OuterVolumeSpecName: "kube-api-access-vgdjw") pod "ce5052d2-0875-421c-8f43-e979da7f8681" (UID: "ce5052d2-0875-421c-8f43-e979da7f8681"). InnerVolumeSpecName "kube-api-access-vgdjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:25:22 crc kubenswrapper[4818]: I1203 07:25:22.970292 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce5052d2-0875-421c-8f43-e979da7f8681-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce5052d2-0875-421c-8f43-e979da7f8681" (UID: "ce5052d2-0875-421c-8f43-e979da7f8681"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.014312 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgdjw\" (UniqueName: \"kubernetes.io/projected/ce5052d2-0875-421c-8f43-e979da7f8681-kube-api-access-vgdjw\") on node \"crc\" DevicePath \"\"" Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.014362 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5052d2-0875-421c-8f43-e979da7f8681-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.014377 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5052d2-0875-421c-8f43-e979da7f8681-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.191615 4818 generic.go:334] "Generic (PLEG): container finished" podID="ce5052d2-0875-421c-8f43-e979da7f8681" containerID="dd43fcd4bf76c596ee5566428d60eed3ce0171c970170ce7e51bd0574a87c8f1" exitCode=0 Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.191675 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv8rv" event={"ID":"ce5052d2-0875-421c-8f43-e979da7f8681","Type":"ContainerDied","Data":"dd43fcd4bf76c596ee5566428d60eed3ce0171c970170ce7e51bd0574a87c8f1"} Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.191710 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv8rv" event={"ID":"ce5052d2-0875-421c-8f43-e979da7f8681","Type":"ContainerDied","Data":"d9790df6aa69f96b2a71b26ffa81e2c045e4f115ac17013b1731d640972a26c1"} Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.191754 4818 scope.go:117] "RemoveContainer" containerID="dd43fcd4bf76c596ee5566428d60eed3ce0171c970170ce7e51bd0574a87c8f1" Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.191982 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fv8rv" Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.243448 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fv8rv"] Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.248360 4818 scope.go:117] "RemoveContainer" containerID="1d25c4c4a4e2d5949868eb08e3bd7f32abaede97557ea2775664da1838d5c3e7" Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.253470 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fv8rv"] Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.282842 4818 scope.go:117] "RemoveContainer" containerID="3a78c83f5019e88325de3c4eebe529076210193c6ed781097a5586c0fff93fa5" Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.342129 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lj86j" Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.342170 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lj86j" Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.344398 4818 scope.go:117] "RemoveContainer" containerID="dd43fcd4bf76c596ee5566428d60eed3ce0171c970170ce7e51bd0574a87c8f1" Dec 03 07:25:23 crc kubenswrapper[4818]: E1203 07:25:23.344902 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd43fcd4bf76c596ee5566428d60eed3ce0171c970170ce7e51bd0574a87c8f1\": container with ID starting with dd43fcd4bf76c596ee5566428d60eed3ce0171c970170ce7e51bd0574a87c8f1 not found: ID does not exist" containerID="dd43fcd4bf76c596ee5566428d60eed3ce0171c970170ce7e51bd0574a87c8f1" Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.344970 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd43fcd4bf76c596ee5566428d60eed3ce0171c970170ce7e51bd0574a87c8f1"} err="failed to get container status \"dd43fcd4bf76c596ee5566428d60eed3ce0171c970170ce7e51bd0574a87c8f1\": rpc error: code = NotFound desc = could not find container \"dd43fcd4bf76c596ee5566428d60eed3ce0171c970170ce7e51bd0574a87c8f1\": container with ID starting with dd43fcd4bf76c596ee5566428d60eed3ce0171c970170ce7e51bd0574a87c8f1 not found: ID does not exist" Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.345010 4818 scope.go:117] "RemoveContainer" containerID="1d25c4c4a4e2d5949868eb08e3bd7f32abaede97557ea2775664da1838d5c3e7" Dec 03 07:25:23 crc kubenswrapper[4818]: E1203 07:25:23.345276 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d25c4c4a4e2d5949868eb08e3bd7f32abaede97557ea2775664da1838d5c3e7\": container with ID starting with 1d25c4c4a4e2d5949868eb08e3bd7f32abaede97557ea2775664da1838d5c3e7 not found: ID does not exist" containerID="1d25c4c4a4e2d5949868eb08e3bd7f32abaede97557ea2775664da1838d5c3e7" Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.345318 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d25c4c4a4e2d5949868eb08e3bd7f32abaede97557ea2775664da1838d5c3e7"} err="failed to get container status \"1d25c4c4a4e2d5949868eb08e3bd7f32abaede97557ea2775664da1838d5c3e7\": rpc error: code = NotFound desc = could not find container \"1d25c4c4a4e2d5949868eb08e3bd7f32abaede97557ea2775664da1838d5c3e7\": container with ID starting with 1d25c4c4a4e2d5949868eb08e3bd7f32abaede97557ea2775664da1838d5c3e7 not found: ID does not exist" Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.345343 4818 scope.go:117] "RemoveContainer" containerID="3a78c83f5019e88325de3c4eebe529076210193c6ed781097a5586c0fff93fa5" Dec 03 07:25:23 crc kubenswrapper[4818]: E1203 07:25:23.345546 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a78c83f5019e88325de3c4eebe529076210193c6ed781097a5586c0fff93fa5\": container with ID starting with 3a78c83f5019e88325de3c4eebe529076210193c6ed781097a5586c0fff93fa5 not found: ID does not exist" containerID="3a78c83f5019e88325de3c4eebe529076210193c6ed781097a5586c0fff93fa5" Dec 03 07:25:23 crc kubenswrapper[4818]: I1203 07:25:23.345584 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a78c83f5019e88325de3c4eebe529076210193c6ed781097a5586c0fff93fa5"} err="failed to get container status \"3a78c83f5019e88325de3c4eebe529076210193c6ed781097a5586c0fff93fa5\": rpc error: code = NotFound desc = could not find container \"3a78c83f5019e88325de3c4eebe529076210193c6ed781097a5586c0fff93fa5\": container with ID starting with 3a78c83f5019e88325de3c4eebe529076210193c6ed781097a5586c0fff93fa5 not found: ID does not exist" Dec 03 07:25:24 crc kubenswrapper[4818]: I1203 07:25:24.395013 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lj86j" podUID="8d18581c-b5a3-4132-afcc-71bced63091b" containerName="registry-server" probeResult="failure" output=< Dec 03 07:25:24 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Dec 03 07:25:24 crc kubenswrapper[4818]: > Dec 03 07:25:24 crc kubenswrapper[4818]: I1203 07:25:24.748216 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce5052d2-0875-421c-8f43-e979da7f8681" path="/var/lib/kubelet/pods/ce5052d2-0875-421c-8f43-e979da7f8681/volumes" Dec 03 07:25:33 crc kubenswrapper[4818]: I1203 07:25:33.412499 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lj86j" Dec 03 07:25:33 crc kubenswrapper[4818]: I1203 07:25:33.477987 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lj86j" Dec 03 07:25:33 crc kubenswrapper[4818]: I1203 07:25:33.656659 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lj86j"] Dec 03 07:25:35 crc kubenswrapper[4818]: I1203 07:25:35.323888 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lj86j" podUID="8d18581c-b5a3-4132-afcc-71bced63091b" containerName="registry-server" containerID="cri-o://e9f0ec583ce4c81e4e42d47520b0d800a19a03bf38c8c27c4ad385abd7203cf9" gracePeriod=2 Dec 03 07:25:35 crc kubenswrapper[4818]: I1203 07:25:35.869435 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lj86j" Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:35.999977 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d18581c-b5a3-4132-afcc-71bced63091b-utilities\") pod \"8d18581c-b5a3-4132-afcc-71bced63091b\" (UID: \"8d18581c-b5a3-4132-afcc-71bced63091b\") " Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.000025 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d18581c-b5a3-4132-afcc-71bced63091b-catalog-content\") pod \"8d18581c-b5a3-4132-afcc-71bced63091b\" (UID: \"8d18581c-b5a3-4132-afcc-71bced63091b\") " Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.000154 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z62gn\" (UniqueName: \"kubernetes.io/projected/8d18581c-b5a3-4132-afcc-71bced63091b-kube-api-access-z62gn\") pod \"8d18581c-b5a3-4132-afcc-71bced63091b\" (UID: \"8d18581c-b5a3-4132-afcc-71bced63091b\") " Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.001060 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d18581c-b5a3-4132-afcc-71bced63091b-utilities" (OuterVolumeSpecName: "utilities") pod "8d18581c-b5a3-4132-afcc-71bced63091b" (UID: "8d18581c-b5a3-4132-afcc-71bced63091b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.002599 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d18581c-b5a3-4132-afcc-71bced63091b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.016140 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d18581c-b5a3-4132-afcc-71bced63091b-kube-api-access-z62gn" (OuterVolumeSpecName: "kube-api-access-z62gn") pod "8d18581c-b5a3-4132-afcc-71bced63091b" (UID: "8d18581c-b5a3-4132-afcc-71bced63091b"). InnerVolumeSpecName "kube-api-access-z62gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.104349 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z62gn\" (UniqueName: \"kubernetes.io/projected/8d18581c-b5a3-4132-afcc-71bced63091b-kube-api-access-z62gn\") on node \"crc\" DevicePath \"\"" Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.104485 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d18581c-b5a3-4132-afcc-71bced63091b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d18581c-b5a3-4132-afcc-71bced63091b" (UID: "8d18581c-b5a3-4132-afcc-71bced63091b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.205918 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d18581c-b5a3-4132-afcc-71bced63091b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.345115 4818 generic.go:334] "Generic (PLEG): container finished" podID="8d18581c-b5a3-4132-afcc-71bced63091b" containerID="e9f0ec583ce4c81e4e42d47520b0d800a19a03bf38c8c27c4ad385abd7203cf9" exitCode=0 Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.346454 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lj86j" Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.346499 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lj86j" event={"ID":"8d18581c-b5a3-4132-afcc-71bced63091b","Type":"ContainerDied","Data":"e9f0ec583ce4c81e4e42d47520b0d800a19a03bf38c8c27c4ad385abd7203cf9"} Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.346860 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lj86j" event={"ID":"8d18581c-b5a3-4132-afcc-71bced63091b","Type":"ContainerDied","Data":"d7ae9c2cb2367c086f4ef3a2299b8ed379b86254cd2b52d529ca067a42af7646"} Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.346929 4818 scope.go:117] "RemoveContainer" containerID="e9f0ec583ce4c81e4e42d47520b0d800a19a03bf38c8c27c4ad385abd7203cf9" Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.385498 4818 scope.go:117] "RemoveContainer" containerID="b90558580c10f4ba9df13dd489b076af606c0623ef2e525fe0f69d711f25de3b" Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.420644 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lj86j"] Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.434383 4818 scope.go:117] "RemoveContainer" containerID="0d465c3eac67ddd1fe0b084045b20a8af4746b13cf450d24240881bd36cdf74a" Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.438529 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lj86j"] Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.476910 4818 scope.go:117] "RemoveContainer" containerID="e9f0ec583ce4c81e4e42d47520b0d800a19a03bf38c8c27c4ad385abd7203cf9" Dec 03 07:25:36 crc kubenswrapper[4818]: E1203 07:25:36.477361 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9f0ec583ce4c81e4e42d47520b0d800a19a03bf38c8c27c4ad385abd7203cf9\": container with ID starting with e9f0ec583ce4c81e4e42d47520b0d800a19a03bf38c8c27c4ad385abd7203cf9 not found: ID does not exist" containerID="e9f0ec583ce4c81e4e42d47520b0d800a19a03bf38c8c27c4ad385abd7203cf9" Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.477407 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9f0ec583ce4c81e4e42d47520b0d800a19a03bf38c8c27c4ad385abd7203cf9"} err="failed to get container status \"e9f0ec583ce4c81e4e42d47520b0d800a19a03bf38c8c27c4ad385abd7203cf9\": rpc error: code = NotFound desc = could not find container \"e9f0ec583ce4c81e4e42d47520b0d800a19a03bf38c8c27c4ad385abd7203cf9\": container with ID starting with e9f0ec583ce4c81e4e42d47520b0d800a19a03bf38c8c27c4ad385abd7203cf9 not found: ID does not exist" Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.477441 4818 scope.go:117] "RemoveContainer" containerID="b90558580c10f4ba9df13dd489b076af606c0623ef2e525fe0f69d711f25de3b" Dec 03 07:25:36 crc kubenswrapper[4818]: E1203 07:25:36.477847 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b90558580c10f4ba9df13dd489b076af606c0623ef2e525fe0f69d711f25de3b\": container with ID starting with b90558580c10f4ba9df13dd489b076af606c0623ef2e525fe0f69d711f25de3b not found: ID does not exist" containerID="b90558580c10f4ba9df13dd489b076af606c0623ef2e525fe0f69d711f25de3b" Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.477878 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b90558580c10f4ba9df13dd489b076af606c0623ef2e525fe0f69d711f25de3b"} err="failed to get container status \"b90558580c10f4ba9df13dd489b076af606c0623ef2e525fe0f69d711f25de3b\": rpc error: code = NotFound desc = could not find container \"b90558580c10f4ba9df13dd489b076af606c0623ef2e525fe0f69d711f25de3b\": container with ID starting with b90558580c10f4ba9df13dd489b076af606c0623ef2e525fe0f69d711f25de3b not found: ID does not exist" Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.477898 4818 scope.go:117] "RemoveContainer" containerID="0d465c3eac67ddd1fe0b084045b20a8af4746b13cf450d24240881bd36cdf74a" Dec 03 07:25:36 crc kubenswrapper[4818]: E1203 07:25:36.478258 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d465c3eac67ddd1fe0b084045b20a8af4746b13cf450d24240881bd36cdf74a\": container with ID starting with 0d465c3eac67ddd1fe0b084045b20a8af4746b13cf450d24240881bd36cdf74a not found: ID does not exist" containerID="0d465c3eac67ddd1fe0b084045b20a8af4746b13cf450d24240881bd36cdf74a" Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.478281 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d465c3eac67ddd1fe0b084045b20a8af4746b13cf450d24240881bd36cdf74a"} err="failed to get container status \"0d465c3eac67ddd1fe0b084045b20a8af4746b13cf450d24240881bd36cdf74a\": rpc error: code = NotFound desc = could not find container \"0d465c3eac67ddd1fe0b084045b20a8af4746b13cf450d24240881bd36cdf74a\": container with ID starting with 0d465c3eac67ddd1fe0b084045b20a8af4746b13cf450d24240881bd36cdf74a not found: ID does not exist" Dec 03 07:25:36 crc kubenswrapper[4818]: I1203 07:25:36.757786 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d18581c-b5a3-4132-afcc-71bced63091b" path="/var/lib/kubelet/pods/8d18581c-b5a3-4132-afcc-71bced63091b/volumes" Dec 03 07:26:43 crc kubenswrapper[4818]: I1203 07:26:43.302196 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:26:43 crc kubenswrapper[4818]: I1203 07:26:43.302747 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.729676 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4lz5h"] Dec 03 07:26:44 crc kubenswrapper[4818]: E1203 07:26:44.730572 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5052d2-0875-421c-8f43-e979da7f8681" containerName="extract-utilities" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.730589 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5052d2-0875-421c-8f43-e979da7f8681" containerName="extract-utilities" Dec 03 07:26:44 crc kubenswrapper[4818]: E1203 07:26:44.730607 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5052d2-0875-421c-8f43-e979da7f8681" containerName="extract-content" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.730615 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5052d2-0875-421c-8f43-e979da7f8681" containerName="extract-content" Dec 03 07:26:44 crc kubenswrapper[4818]: E1203 07:26:44.730629 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d18581c-b5a3-4132-afcc-71bced63091b" containerName="extract-content" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.730637 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d18581c-b5a3-4132-afcc-71bced63091b" containerName="extract-content" Dec 03 07:26:44 crc kubenswrapper[4818]: E1203 07:26:44.730653 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d18581c-b5a3-4132-afcc-71bced63091b" containerName="registry-server" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.730661 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d18581c-b5a3-4132-afcc-71bced63091b" containerName="registry-server" Dec 03 07:26:44 crc kubenswrapper[4818]: E1203 07:26:44.730697 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5052d2-0875-421c-8f43-e979da7f8681" containerName="registry-server" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.730704 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5052d2-0875-421c-8f43-e979da7f8681" containerName="registry-server" Dec 03 07:26:44 crc kubenswrapper[4818]: E1203 07:26:44.730721 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d18581c-b5a3-4132-afcc-71bced63091b" containerName="extract-utilities" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.730728 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d18581c-b5a3-4132-afcc-71bced63091b" containerName="extract-utilities" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.730993 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce5052d2-0875-421c-8f43-e979da7f8681" containerName="registry-server" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.731017 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d18581c-b5a3-4132-afcc-71bced63091b" containerName="registry-server" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.732798 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4lz5h" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.771147 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4lz5h"] Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.858750 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd57b298-5e38-4896-a2e6-b5206865b79c-catalog-content\") pod \"community-operators-4lz5h\" (UID: \"bd57b298-5e38-4896-a2e6-b5206865b79c\") " pod="openshift-marketplace/community-operators-4lz5h" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.859179 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99rjh\" (UniqueName: \"kubernetes.io/projected/bd57b298-5e38-4896-a2e6-b5206865b79c-kube-api-access-99rjh\") pod \"community-operators-4lz5h\" (UID: \"bd57b298-5e38-4896-a2e6-b5206865b79c\") " pod="openshift-marketplace/community-operators-4lz5h" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.859278 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd57b298-5e38-4896-a2e6-b5206865b79c-utilities\") pod \"community-operators-4lz5h\" (UID: \"bd57b298-5e38-4896-a2e6-b5206865b79c\") " pod="openshift-marketplace/community-operators-4lz5h" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.960912 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99rjh\" (UniqueName: \"kubernetes.io/projected/bd57b298-5e38-4896-a2e6-b5206865b79c-kube-api-access-99rjh\") pod \"community-operators-4lz5h\" (UID: \"bd57b298-5e38-4896-a2e6-b5206865b79c\") " pod="openshift-marketplace/community-operators-4lz5h" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.961072 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd57b298-5e38-4896-a2e6-b5206865b79c-utilities\") pod \"community-operators-4lz5h\" (UID: \"bd57b298-5e38-4896-a2e6-b5206865b79c\") " pod="openshift-marketplace/community-operators-4lz5h" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.961129 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd57b298-5e38-4896-a2e6-b5206865b79c-catalog-content\") pod \"community-operators-4lz5h\" (UID: \"bd57b298-5e38-4896-a2e6-b5206865b79c\") " pod="openshift-marketplace/community-operators-4lz5h" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.961686 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd57b298-5e38-4896-a2e6-b5206865b79c-utilities\") pod \"community-operators-4lz5h\" (UID: \"bd57b298-5e38-4896-a2e6-b5206865b79c\") " pod="openshift-marketplace/community-operators-4lz5h" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.961714 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd57b298-5e38-4896-a2e6-b5206865b79c-catalog-content\") pod \"community-operators-4lz5h\" (UID: \"bd57b298-5e38-4896-a2e6-b5206865b79c\") " pod="openshift-marketplace/community-operators-4lz5h" Dec 03 07:26:44 crc kubenswrapper[4818]: I1203 07:26:44.979300 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99rjh\" (UniqueName: \"kubernetes.io/projected/bd57b298-5e38-4896-a2e6-b5206865b79c-kube-api-access-99rjh\") pod \"community-operators-4lz5h\" (UID: \"bd57b298-5e38-4896-a2e6-b5206865b79c\") " pod="openshift-marketplace/community-operators-4lz5h" Dec 03 07:26:45 crc kubenswrapper[4818]: I1203 07:26:45.104865 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4lz5h" Dec 03 07:26:45 crc kubenswrapper[4818]: I1203 07:26:45.606945 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4lz5h"] Dec 03 07:26:46 crc kubenswrapper[4818]: I1203 07:26:46.085705 4818 generic.go:334] "Generic (PLEG): container finished" podID="bd57b298-5e38-4896-a2e6-b5206865b79c" containerID="277388b4c8553b1a0521f89e9b9b07f1f492acdad80ad8e2e9acb2a237f91773" exitCode=0 Dec 03 07:26:46 crc kubenswrapper[4818]: I1203 07:26:46.085782 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4lz5h" event={"ID":"bd57b298-5e38-4896-a2e6-b5206865b79c","Type":"ContainerDied","Data":"277388b4c8553b1a0521f89e9b9b07f1f492acdad80ad8e2e9acb2a237f91773"} Dec 03 07:26:46 crc kubenswrapper[4818]: I1203 07:26:46.086967 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4lz5h" event={"ID":"bd57b298-5e38-4896-a2e6-b5206865b79c","Type":"ContainerStarted","Data":"ff59043211b57619aafbcc3fe2688b6160414ffa2a6cfb4c5e174bbe166d1ead"} Dec 03 07:26:47 crc kubenswrapper[4818]: I1203 07:26:47.098226 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4lz5h" event={"ID":"bd57b298-5e38-4896-a2e6-b5206865b79c","Type":"ContainerStarted","Data":"a074d33d87aa6c3271f6c2e5396245b6c7a8ecc17b397103136e1688f6aeb8b8"} Dec 03 07:26:48 crc kubenswrapper[4818]: I1203 07:26:48.113143 4818 generic.go:334] "Generic (PLEG): container finished" podID="bd57b298-5e38-4896-a2e6-b5206865b79c" containerID="a074d33d87aa6c3271f6c2e5396245b6c7a8ecc17b397103136e1688f6aeb8b8" exitCode=0 Dec 03 07:26:48 crc kubenswrapper[4818]: I1203 07:26:48.113255 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4lz5h" event={"ID":"bd57b298-5e38-4896-a2e6-b5206865b79c","Type":"ContainerDied","Data":"a074d33d87aa6c3271f6c2e5396245b6c7a8ecc17b397103136e1688f6aeb8b8"} Dec 03 07:26:49 crc kubenswrapper[4818]: I1203 07:26:49.128636 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4lz5h" event={"ID":"bd57b298-5e38-4896-a2e6-b5206865b79c","Type":"ContainerStarted","Data":"c9550c13f0a9e9b8208d19faf99260edec74effea4a93d961a501ac5aa27fdcd"} Dec 03 07:26:49 crc kubenswrapper[4818]: I1203 07:26:49.156565 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4lz5h" podStartSLOduration=2.72858427 podStartE2EDuration="5.156545141s" podCreationTimestamp="2025-12-03 07:26:44 +0000 UTC" firstStartedPulling="2025-12-03 07:26:46.087775188 +0000 UTC m=+3563.779383950" lastFinishedPulling="2025-12-03 07:26:48.515736059 +0000 UTC m=+3566.207344821" observedRunningTime="2025-12-03 07:26:49.150255254 +0000 UTC m=+3566.841864016" watchObservedRunningTime="2025-12-03 07:26:49.156545141 +0000 UTC m=+3566.848153893" Dec 03 07:26:55 crc kubenswrapper[4818]: I1203 07:26:55.105535 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4lz5h" Dec 03 07:26:55 crc kubenswrapper[4818]: I1203 07:26:55.106200 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4lz5h" Dec 03 07:26:55 crc kubenswrapper[4818]: I1203 07:26:55.191373 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4lz5h" Dec 03 07:26:55 crc kubenswrapper[4818]: I1203 07:26:55.343413 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4lz5h" Dec 03 07:26:55 crc kubenswrapper[4818]: I1203 07:26:55.446310 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4lz5h"] Dec 03 07:26:57 crc kubenswrapper[4818]: I1203 07:26:57.236708 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4lz5h" podUID="bd57b298-5e38-4896-a2e6-b5206865b79c" containerName="registry-server" containerID="cri-o://c9550c13f0a9e9b8208d19faf99260edec74effea4a93d961a501ac5aa27fdcd" gracePeriod=2 Dec 03 07:26:58 crc kubenswrapper[4818]: I1203 07:26:58.250957 4818 generic.go:334] "Generic (PLEG): container finished" podID="bd57b298-5e38-4896-a2e6-b5206865b79c" containerID="c9550c13f0a9e9b8208d19faf99260edec74effea4a93d961a501ac5aa27fdcd" exitCode=0 Dec 03 07:26:58 crc kubenswrapper[4818]: I1203 07:26:58.251276 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4lz5h" event={"ID":"bd57b298-5e38-4896-a2e6-b5206865b79c","Type":"ContainerDied","Data":"c9550c13f0a9e9b8208d19faf99260edec74effea4a93d961a501ac5aa27fdcd"} Dec 03 07:26:58 crc kubenswrapper[4818]: I1203 07:26:58.365002 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4lz5h" Dec 03 07:26:58 crc kubenswrapper[4818]: I1203 07:26:58.489419 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd57b298-5e38-4896-a2e6-b5206865b79c-catalog-content\") pod \"bd57b298-5e38-4896-a2e6-b5206865b79c\" (UID: \"bd57b298-5e38-4896-a2e6-b5206865b79c\") " Dec 03 07:26:58 crc kubenswrapper[4818]: I1203 07:26:58.489609 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99rjh\" (UniqueName: \"kubernetes.io/projected/bd57b298-5e38-4896-a2e6-b5206865b79c-kube-api-access-99rjh\") pod \"bd57b298-5e38-4896-a2e6-b5206865b79c\" (UID: \"bd57b298-5e38-4896-a2e6-b5206865b79c\") " Dec 03 07:26:58 crc kubenswrapper[4818]: I1203 07:26:58.491067 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd57b298-5e38-4896-a2e6-b5206865b79c-utilities\") pod \"bd57b298-5e38-4896-a2e6-b5206865b79c\" (UID: \"bd57b298-5e38-4896-a2e6-b5206865b79c\") " Dec 03 07:26:58 crc kubenswrapper[4818]: I1203 07:26:58.492791 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd57b298-5e38-4896-a2e6-b5206865b79c-utilities" (OuterVolumeSpecName: "utilities") pod "bd57b298-5e38-4896-a2e6-b5206865b79c" (UID: "bd57b298-5e38-4896-a2e6-b5206865b79c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:26:58 crc kubenswrapper[4818]: I1203 07:26:58.496377 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd57b298-5e38-4896-a2e6-b5206865b79c-kube-api-access-99rjh" (OuterVolumeSpecName: "kube-api-access-99rjh") pod "bd57b298-5e38-4896-a2e6-b5206865b79c" (UID: "bd57b298-5e38-4896-a2e6-b5206865b79c"). InnerVolumeSpecName "kube-api-access-99rjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:26:58 crc kubenswrapper[4818]: I1203 07:26:58.546578 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd57b298-5e38-4896-a2e6-b5206865b79c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bd57b298-5e38-4896-a2e6-b5206865b79c" (UID: "bd57b298-5e38-4896-a2e6-b5206865b79c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:26:58 crc kubenswrapper[4818]: I1203 07:26:58.593019 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd57b298-5e38-4896-a2e6-b5206865b79c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:26:58 crc kubenswrapper[4818]: I1203 07:26:58.593053 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd57b298-5e38-4896-a2e6-b5206865b79c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:26:58 crc kubenswrapper[4818]: I1203 07:26:58.593063 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99rjh\" (UniqueName: \"kubernetes.io/projected/bd57b298-5e38-4896-a2e6-b5206865b79c-kube-api-access-99rjh\") on node \"crc\" DevicePath \"\"" Dec 03 07:26:59 crc kubenswrapper[4818]: I1203 07:26:59.263350 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4lz5h" event={"ID":"bd57b298-5e38-4896-a2e6-b5206865b79c","Type":"ContainerDied","Data":"ff59043211b57619aafbcc3fe2688b6160414ffa2a6cfb4c5e174bbe166d1ead"} Dec 03 07:26:59 crc kubenswrapper[4818]: I1203 07:26:59.263438 4818 scope.go:117] "RemoveContainer" containerID="c9550c13f0a9e9b8208d19faf99260edec74effea4a93d961a501ac5aa27fdcd" Dec 03 07:26:59 crc kubenswrapper[4818]: I1203 07:26:59.263543 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4lz5h" Dec 03 07:26:59 crc kubenswrapper[4818]: I1203 07:26:59.294287 4818 scope.go:117] "RemoveContainer" containerID="a074d33d87aa6c3271f6c2e5396245b6c7a8ecc17b397103136e1688f6aeb8b8" Dec 03 07:26:59 crc kubenswrapper[4818]: I1203 07:26:59.295580 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4lz5h"] Dec 03 07:26:59 crc kubenswrapper[4818]: I1203 07:26:59.303308 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4lz5h"] Dec 03 07:26:59 crc kubenswrapper[4818]: I1203 07:26:59.319770 4818 scope.go:117] "RemoveContainer" containerID="277388b4c8553b1a0521f89e9b9b07f1f492acdad80ad8e2e9acb2a237f91773" Dec 03 07:27:00 crc kubenswrapper[4818]: I1203 07:27:00.747624 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd57b298-5e38-4896-a2e6-b5206865b79c" path="/var/lib/kubelet/pods/bd57b298-5e38-4896-a2e6-b5206865b79c/volumes" Dec 03 07:27:10 crc kubenswrapper[4818]: I1203 07:27:10.361774 4818 generic.go:334] "Generic (PLEG): container finished" podID="a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf" containerID="cd6cbaba39d8cfd15e18267f59419e26bc7d2c41b6b936f55775ba3ebddd71c4" exitCode=0 Dec 03 07:27:10 crc kubenswrapper[4818]: I1203 07:27:10.361881 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf","Type":"ContainerDied","Data":"cd6cbaba39d8cfd15e18267f59419e26bc7d2c41b6b936f55775ba3ebddd71c4"} Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.778342 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.891244 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-ca-certs\") pod \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.892307 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-openstack-config\") pod \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.892354 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-ssh-key\") pod \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.892441 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.892528 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-test-operator-ephemeral-temporary\") pod \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.892602 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-test-operator-ephemeral-workdir\") pod \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.892680 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxjcz\" (UniqueName: \"kubernetes.io/projected/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-kube-api-access-vxjcz\") pod \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.892712 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-config-data\") pod \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.892762 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-openstack-config-secret\") pod \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\" (UID: \"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf\") " Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.893283 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf" (UID: "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.894280 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-config-data" (OuterVolumeSpecName: "config-data") pod "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf" (UID: "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.898074 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf" (UID: "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.898221 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-kube-api-access-vxjcz" (OuterVolumeSpecName: "kube-api-access-vxjcz") pod "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf" (UID: "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf"). InnerVolumeSpecName "kube-api-access-vxjcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.904733 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf" (UID: "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.927420 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf" (UID: "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.928371 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf" (UID: "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.945016 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf" (UID: "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.958534 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf" (UID: "a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.995368 4818 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.995399 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxjcz\" (UniqueName: \"kubernetes.io/projected/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-kube-api-access-vxjcz\") on node \"crc\" DevicePath \"\"" Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.995411 4818 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.995420 4818 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.995447 4818 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.995456 4818 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.995463 4818 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.995504 4818 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 03 07:27:11 crc kubenswrapper[4818]: I1203 07:27:11.995514 4818 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 03 07:27:12 crc kubenswrapper[4818]: I1203 07:27:12.017318 4818 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 03 07:27:12 crc kubenswrapper[4818]: I1203 07:27:12.098042 4818 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 03 07:27:12 crc kubenswrapper[4818]: I1203 07:27:12.398414 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf","Type":"ContainerDied","Data":"4f25090c7c16f6d3a82b065182ed88abec3b200807a6c5b7588126d0ea5a1870"} Dec 03 07:27:12 crc kubenswrapper[4818]: I1203 07:27:12.398454 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f25090c7c16f6d3a82b065182ed88abec3b200807a6c5b7588126d0ea5a1870" Dec 03 07:27:12 crc kubenswrapper[4818]: I1203 07:27:12.398474 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 07:27:13 crc kubenswrapper[4818]: I1203 07:27:13.302961 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:27:13 crc kubenswrapper[4818]: I1203 07:27:13.303282 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:27:14 crc kubenswrapper[4818]: I1203 07:27:14.967216 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 07:27:14 crc kubenswrapper[4818]: E1203 07:27:14.967686 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd57b298-5e38-4896-a2e6-b5206865b79c" containerName="registry-server" Dec 03 07:27:14 crc kubenswrapper[4818]: I1203 07:27:14.967702 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd57b298-5e38-4896-a2e6-b5206865b79c" containerName="registry-server" Dec 03 07:27:14 crc kubenswrapper[4818]: E1203 07:27:14.967716 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd57b298-5e38-4896-a2e6-b5206865b79c" containerName="extract-utilities" Dec 03 07:27:14 crc kubenswrapper[4818]: I1203 07:27:14.967727 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd57b298-5e38-4896-a2e6-b5206865b79c" containerName="extract-utilities" Dec 03 07:27:14 crc kubenswrapper[4818]: E1203 07:27:14.967762 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf" containerName="tempest-tests-tempest-tests-runner" Dec 03 07:27:14 crc kubenswrapper[4818]: I1203 07:27:14.967771 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf" containerName="tempest-tests-tempest-tests-runner" Dec 03 07:27:14 crc kubenswrapper[4818]: E1203 07:27:14.967785 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd57b298-5e38-4896-a2e6-b5206865b79c" containerName="extract-content" Dec 03 07:27:14 crc kubenswrapper[4818]: I1203 07:27:14.967793 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd57b298-5e38-4896-a2e6-b5206865b79c" containerName="extract-content" Dec 03 07:27:14 crc kubenswrapper[4818]: I1203 07:27:14.968067 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf" containerName="tempest-tests-tempest-tests-runner" Dec 03 07:27:14 crc kubenswrapper[4818]: I1203 07:27:14.968089 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd57b298-5e38-4896-a2e6-b5206865b79c" containerName="registry-server" Dec 03 07:27:14 crc kubenswrapper[4818]: I1203 07:27:14.968910 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 07:27:14 crc kubenswrapper[4818]: I1203 07:27:14.973534 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-rwqql" Dec 03 07:27:14 crc kubenswrapper[4818]: I1203 07:27:14.982526 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 07:27:15 crc kubenswrapper[4818]: I1203 07:27:15.059992 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl8pt\" (UniqueName: \"kubernetes.io/projected/78a39ae7-b979-4d32-a128-25dca0a393e4-kube-api-access-pl8pt\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"78a39ae7-b979-4d32-a128-25dca0a393e4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 07:27:15 crc kubenswrapper[4818]: I1203 07:27:15.060477 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"78a39ae7-b979-4d32-a128-25dca0a393e4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 07:27:15 crc kubenswrapper[4818]: I1203 07:27:15.162272 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl8pt\" (UniqueName: \"kubernetes.io/projected/78a39ae7-b979-4d32-a128-25dca0a393e4-kube-api-access-pl8pt\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"78a39ae7-b979-4d32-a128-25dca0a393e4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 07:27:15 crc kubenswrapper[4818]: I1203 07:27:15.162725 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"78a39ae7-b979-4d32-a128-25dca0a393e4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 07:27:15 crc kubenswrapper[4818]: I1203 07:27:15.163114 4818 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"78a39ae7-b979-4d32-a128-25dca0a393e4\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 07:27:15 crc kubenswrapper[4818]: I1203 07:27:15.195194 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl8pt\" (UniqueName: \"kubernetes.io/projected/78a39ae7-b979-4d32-a128-25dca0a393e4-kube-api-access-pl8pt\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"78a39ae7-b979-4d32-a128-25dca0a393e4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 07:27:15 crc kubenswrapper[4818]: I1203 07:27:15.197325 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"78a39ae7-b979-4d32-a128-25dca0a393e4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 07:27:15 crc kubenswrapper[4818]: I1203 07:27:15.288497 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 07:27:15 crc kubenswrapper[4818]: I1203 07:27:15.770841 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 07:27:16 crc kubenswrapper[4818]: I1203 07:27:16.439687 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"78a39ae7-b979-4d32-a128-25dca0a393e4","Type":"ContainerStarted","Data":"b1a319e267b269c761f502e33b10f9d605f427a6d52f5a5cec70e59c4fd8ad5e"} Dec 03 07:27:17 crc kubenswrapper[4818]: I1203 07:27:17.450195 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"78a39ae7-b979-4d32-a128-25dca0a393e4","Type":"ContainerStarted","Data":"7666b5d471c96b076661de870a76ac750dfd173f875a65aba147231063999bbd"} Dec 03 07:27:17 crc kubenswrapper[4818]: I1203 07:27:17.466412 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.719679911 podStartE2EDuration="3.466392233s" podCreationTimestamp="2025-12-03 07:27:14 +0000 UTC" firstStartedPulling="2025-12-03 07:27:15.773467913 +0000 UTC m=+3593.465076665" lastFinishedPulling="2025-12-03 07:27:16.520180235 +0000 UTC m=+3594.211788987" observedRunningTime="2025-12-03 07:27:17.464183648 +0000 UTC m=+3595.155792420" watchObservedRunningTime="2025-12-03 07:27:17.466392233 +0000 UTC m=+3595.158000985" Dec 03 07:27:39 crc kubenswrapper[4818]: I1203 07:27:39.861201 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ffrv8/must-gather-2cb84"] Dec 03 07:27:39 crc kubenswrapper[4818]: I1203 07:27:39.863423 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffrv8/must-gather-2cb84" Dec 03 07:27:39 crc kubenswrapper[4818]: I1203 07:27:39.865875 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ffrv8"/"openshift-service-ca.crt" Dec 03 07:27:39 crc kubenswrapper[4818]: I1203 07:27:39.867639 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ffrv8"/"kube-root-ca.crt" Dec 03 07:27:39 crc kubenswrapper[4818]: I1203 07:27:39.880748 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ffrv8/must-gather-2cb84"] Dec 03 07:27:40 crc kubenswrapper[4818]: I1203 07:27:40.053675 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/18e6fb67-550a-4057-a619-6ccfb3707649-must-gather-output\") pod \"must-gather-2cb84\" (UID: \"18e6fb67-550a-4057-a619-6ccfb3707649\") " pod="openshift-must-gather-ffrv8/must-gather-2cb84" Dec 03 07:27:40 crc kubenswrapper[4818]: I1203 07:27:40.053855 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4v5m\" (UniqueName: \"kubernetes.io/projected/18e6fb67-550a-4057-a619-6ccfb3707649-kube-api-access-l4v5m\") pod \"must-gather-2cb84\" (UID: \"18e6fb67-550a-4057-a619-6ccfb3707649\") " pod="openshift-must-gather-ffrv8/must-gather-2cb84" Dec 03 07:27:40 crc kubenswrapper[4818]: I1203 07:27:40.156862 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/18e6fb67-550a-4057-a619-6ccfb3707649-must-gather-output\") pod \"must-gather-2cb84\" (UID: \"18e6fb67-550a-4057-a619-6ccfb3707649\") " pod="openshift-must-gather-ffrv8/must-gather-2cb84" Dec 03 07:27:40 crc kubenswrapper[4818]: I1203 07:27:40.157044 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4v5m\" (UniqueName: \"kubernetes.io/projected/18e6fb67-550a-4057-a619-6ccfb3707649-kube-api-access-l4v5m\") pod \"must-gather-2cb84\" (UID: \"18e6fb67-550a-4057-a619-6ccfb3707649\") " pod="openshift-must-gather-ffrv8/must-gather-2cb84" Dec 03 07:27:40 crc kubenswrapper[4818]: I1203 07:27:40.157399 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/18e6fb67-550a-4057-a619-6ccfb3707649-must-gather-output\") pod \"must-gather-2cb84\" (UID: \"18e6fb67-550a-4057-a619-6ccfb3707649\") " pod="openshift-must-gather-ffrv8/must-gather-2cb84" Dec 03 07:27:40 crc kubenswrapper[4818]: I1203 07:27:40.175646 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4v5m\" (UniqueName: \"kubernetes.io/projected/18e6fb67-550a-4057-a619-6ccfb3707649-kube-api-access-l4v5m\") pod \"must-gather-2cb84\" (UID: \"18e6fb67-550a-4057-a619-6ccfb3707649\") " pod="openshift-must-gather-ffrv8/must-gather-2cb84" Dec 03 07:27:40 crc kubenswrapper[4818]: I1203 07:27:40.193293 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffrv8/must-gather-2cb84" Dec 03 07:27:40 crc kubenswrapper[4818]: I1203 07:27:40.645173 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ffrv8/must-gather-2cb84"] Dec 03 07:27:40 crc kubenswrapper[4818]: W1203 07:27:40.646909 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18e6fb67_550a_4057_a619_6ccfb3707649.slice/crio-341f33b3d92f50c0457ef3529ba11ef893e0733daedacac87a0a693eb4a061f9 WatchSource:0}: Error finding container 341f33b3d92f50c0457ef3529ba11ef893e0733daedacac87a0a693eb4a061f9: Status 404 returned error can't find the container with id 341f33b3d92f50c0457ef3529ba11ef893e0733daedacac87a0a693eb4a061f9 Dec 03 07:27:40 crc kubenswrapper[4818]: I1203 07:27:40.725717 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffrv8/must-gather-2cb84" event={"ID":"18e6fb67-550a-4057-a619-6ccfb3707649","Type":"ContainerStarted","Data":"341f33b3d92f50c0457ef3529ba11ef893e0733daedacac87a0a693eb4a061f9"} Dec 03 07:27:43 crc kubenswrapper[4818]: I1203 07:27:43.301974 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:27:43 crc kubenswrapper[4818]: I1203 07:27:43.302283 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:27:43 crc kubenswrapper[4818]: I1203 07:27:43.302325 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 07:27:43 crc kubenswrapper[4818]: I1203 07:27:43.303112 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020"} pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:27:43 crc kubenswrapper[4818]: I1203 07:27:43.303166 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" containerID="cri-o://e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" gracePeriod=600 Dec 03 07:27:43 crc kubenswrapper[4818]: I1203 07:27:43.780979 4818 generic.go:334] "Generic (PLEG): container finished" podID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" exitCode=0 Dec 03 07:27:43 crc kubenswrapper[4818]: I1203 07:27:43.781035 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerDied","Data":"e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020"} Dec 03 07:27:43 crc kubenswrapper[4818]: I1203 07:27:43.781553 4818 scope.go:117] "RemoveContainer" containerID="fe2c3fb698cd81e0da40cc4153b5a00fd512ba89b40ba70448f8ea1e0ace6325" Dec 03 07:27:44 crc kubenswrapper[4818]: E1203 07:27:44.266397 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:27:44 crc kubenswrapper[4818]: I1203 07:27:44.793968 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:27:44 crc kubenswrapper[4818]: E1203 07:27:44.794490 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:27:44 crc kubenswrapper[4818]: I1203 07:27:44.796386 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffrv8/must-gather-2cb84" event={"ID":"18e6fb67-550a-4057-a619-6ccfb3707649","Type":"ContainerStarted","Data":"dcefd7bbb59c7b1a55fb0d1e9f927d7a0357766e6cc5bdc543da12480e666e82"} Dec 03 07:27:45 crc kubenswrapper[4818]: I1203 07:27:45.806713 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffrv8/must-gather-2cb84" event={"ID":"18e6fb67-550a-4057-a619-6ccfb3707649","Type":"ContainerStarted","Data":"68c62b78adde405f1777d96a033669e55a0ef61cb60ab62dbefaf335e3513038"} Dec 03 07:27:48 crc kubenswrapper[4818]: I1203 07:27:48.115452 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ffrv8/must-gather-2cb84" podStartSLOduration=5.425573592 podStartE2EDuration="9.115432515s" podCreationTimestamp="2025-12-03 07:27:39 +0000 UTC" firstStartedPulling="2025-12-03 07:27:40.649646432 +0000 UTC m=+3618.341255184" lastFinishedPulling="2025-12-03 07:27:44.339505355 +0000 UTC m=+3622.031114107" observedRunningTime="2025-12-03 07:27:45.837314299 +0000 UTC m=+3623.528923051" watchObservedRunningTime="2025-12-03 07:27:48.115432515 +0000 UTC m=+3625.807041267" Dec 03 07:27:48 crc kubenswrapper[4818]: I1203 07:27:48.122328 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ffrv8/crc-debug-vghqr"] Dec 03 07:27:48 crc kubenswrapper[4818]: I1203 07:27:48.123604 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffrv8/crc-debug-vghqr" Dec 03 07:27:48 crc kubenswrapper[4818]: I1203 07:27:48.125628 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ffrv8"/"default-dockercfg-gw7ft" Dec 03 07:27:48 crc kubenswrapper[4818]: I1203 07:27:48.206664 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a65c921d-45bd-4459-b994-98eb2cf2f9e4-host\") pod \"crc-debug-vghqr\" (UID: \"a65c921d-45bd-4459-b994-98eb2cf2f9e4\") " pod="openshift-must-gather-ffrv8/crc-debug-vghqr" Dec 03 07:27:48 crc kubenswrapper[4818]: I1203 07:27:48.206741 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh8ck\" (UniqueName: \"kubernetes.io/projected/a65c921d-45bd-4459-b994-98eb2cf2f9e4-kube-api-access-kh8ck\") pod \"crc-debug-vghqr\" (UID: \"a65c921d-45bd-4459-b994-98eb2cf2f9e4\") " pod="openshift-must-gather-ffrv8/crc-debug-vghqr" Dec 03 07:27:48 crc kubenswrapper[4818]: I1203 07:27:48.308518 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a65c921d-45bd-4459-b994-98eb2cf2f9e4-host\") pod \"crc-debug-vghqr\" (UID: \"a65c921d-45bd-4459-b994-98eb2cf2f9e4\") " pod="openshift-must-gather-ffrv8/crc-debug-vghqr" Dec 03 07:27:48 crc kubenswrapper[4818]: I1203 07:27:48.308606 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh8ck\" (UniqueName: \"kubernetes.io/projected/a65c921d-45bd-4459-b994-98eb2cf2f9e4-kube-api-access-kh8ck\") pod \"crc-debug-vghqr\" (UID: \"a65c921d-45bd-4459-b994-98eb2cf2f9e4\") " pod="openshift-must-gather-ffrv8/crc-debug-vghqr" Dec 03 07:27:48 crc kubenswrapper[4818]: I1203 07:27:48.308687 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a65c921d-45bd-4459-b994-98eb2cf2f9e4-host\") pod \"crc-debug-vghqr\" (UID: \"a65c921d-45bd-4459-b994-98eb2cf2f9e4\") " pod="openshift-must-gather-ffrv8/crc-debug-vghqr" Dec 03 07:27:48 crc kubenswrapper[4818]: I1203 07:27:48.339274 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh8ck\" (UniqueName: \"kubernetes.io/projected/a65c921d-45bd-4459-b994-98eb2cf2f9e4-kube-api-access-kh8ck\") pod \"crc-debug-vghqr\" (UID: \"a65c921d-45bd-4459-b994-98eb2cf2f9e4\") " pod="openshift-must-gather-ffrv8/crc-debug-vghqr" Dec 03 07:27:48 crc kubenswrapper[4818]: I1203 07:27:48.443755 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffrv8/crc-debug-vghqr" Dec 03 07:27:48 crc kubenswrapper[4818]: W1203 07:27:48.484005 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda65c921d_45bd_4459_b994_98eb2cf2f9e4.slice/crio-7cf09539e3568e2ba392da9da6b3a5b5446cf4491e049f820e23b0aaa9a4bdc3 WatchSource:0}: Error finding container 7cf09539e3568e2ba392da9da6b3a5b5446cf4491e049f820e23b0aaa9a4bdc3: Status 404 returned error can't find the container with id 7cf09539e3568e2ba392da9da6b3a5b5446cf4491e049f820e23b0aaa9a4bdc3 Dec 03 07:27:48 crc kubenswrapper[4818]: I1203 07:27:48.831570 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffrv8/crc-debug-vghqr" event={"ID":"a65c921d-45bd-4459-b994-98eb2cf2f9e4","Type":"ContainerStarted","Data":"7cf09539e3568e2ba392da9da6b3a5b5446cf4491e049f820e23b0aaa9a4bdc3"} Dec 03 07:27:58 crc kubenswrapper[4818]: I1203 07:27:58.738941 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:27:58 crc kubenswrapper[4818]: E1203 07:27:58.739866 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:27:58 crc kubenswrapper[4818]: I1203 07:27:58.949459 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffrv8/crc-debug-vghqr" event={"ID":"a65c921d-45bd-4459-b994-98eb2cf2f9e4","Type":"ContainerStarted","Data":"3ad7210bf67e974d272a456badab288d16e6eb64d4ce8f685ab2c73b20f50f12"} Dec 03 07:27:58 crc kubenswrapper[4818]: I1203 07:27:58.969276 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ffrv8/crc-debug-vghqr" podStartSLOduration=1.097465485 podStartE2EDuration="10.969258402s" podCreationTimestamp="2025-12-03 07:27:48 +0000 UTC" firstStartedPulling="2025-12-03 07:27:48.486440607 +0000 UTC m=+3626.178049369" lastFinishedPulling="2025-12-03 07:27:58.358233534 +0000 UTC m=+3636.049842286" observedRunningTime="2025-12-03 07:27:58.961066038 +0000 UTC m=+3636.652674790" watchObservedRunningTime="2025-12-03 07:27:58.969258402 +0000 UTC m=+3636.660867154" Dec 03 07:28:11 crc kubenswrapper[4818]: I1203 07:28:11.738703 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:28:11 crc kubenswrapper[4818]: E1203 07:28:11.739425 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:28:23 crc kubenswrapper[4818]: I1203 07:28:23.737944 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:28:23 crc kubenswrapper[4818]: E1203 07:28:23.738833 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:28:35 crc kubenswrapper[4818]: I1203 07:28:35.738044 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:28:35 crc kubenswrapper[4818]: E1203 07:28:35.738766 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:28:37 crc kubenswrapper[4818]: I1203 07:28:37.493358 4818 generic.go:334] "Generic (PLEG): container finished" podID="a65c921d-45bd-4459-b994-98eb2cf2f9e4" containerID="3ad7210bf67e974d272a456badab288d16e6eb64d4ce8f685ab2c73b20f50f12" exitCode=0 Dec 03 07:28:37 crc kubenswrapper[4818]: I1203 07:28:37.493704 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffrv8/crc-debug-vghqr" event={"ID":"a65c921d-45bd-4459-b994-98eb2cf2f9e4","Type":"ContainerDied","Data":"3ad7210bf67e974d272a456badab288d16e6eb64d4ce8f685ab2c73b20f50f12"} Dec 03 07:28:38 crc kubenswrapper[4818]: I1203 07:28:38.651630 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffrv8/crc-debug-vghqr" Dec 03 07:28:38 crc kubenswrapper[4818]: I1203 07:28:38.684437 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ffrv8/crc-debug-vghqr"] Dec 03 07:28:38 crc kubenswrapper[4818]: I1203 07:28:38.693342 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ffrv8/crc-debug-vghqr"] Dec 03 07:28:38 crc kubenswrapper[4818]: I1203 07:28:38.767462 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kh8ck\" (UniqueName: \"kubernetes.io/projected/a65c921d-45bd-4459-b994-98eb2cf2f9e4-kube-api-access-kh8ck\") pod \"a65c921d-45bd-4459-b994-98eb2cf2f9e4\" (UID: \"a65c921d-45bd-4459-b994-98eb2cf2f9e4\") " Dec 03 07:28:38 crc kubenswrapper[4818]: I1203 07:28:38.767597 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a65c921d-45bd-4459-b994-98eb2cf2f9e4-host\") pod \"a65c921d-45bd-4459-b994-98eb2cf2f9e4\" (UID: \"a65c921d-45bd-4459-b994-98eb2cf2f9e4\") " Dec 03 07:28:38 crc kubenswrapper[4818]: I1203 07:28:38.767712 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a65c921d-45bd-4459-b994-98eb2cf2f9e4-host" (OuterVolumeSpecName: "host") pod "a65c921d-45bd-4459-b994-98eb2cf2f9e4" (UID: "a65c921d-45bd-4459-b994-98eb2cf2f9e4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:28:38 crc kubenswrapper[4818]: I1203 07:28:38.768806 4818 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a65c921d-45bd-4459-b994-98eb2cf2f9e4-host\") on node \"crc\" DevicePath \"\"" Dec 03 07:28:38 crc kubenswrapper[4818]: I1203 07:28:38.773145 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a65c921d-45bd-4459-b994-98eb2cf2f9e4-kube-api-access-kh8ck" (OuterVolumeSpecName: "kube-api-access-kh8ck") pod "a65c921d-45bd-4459-b994-98eb2cf2f9e4" (UID: "a65c921d-45bd-4459-b994-98eb2cf2f9e4"). InnerVolumeSpecName "kube-api-access-kh8ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:28:38 crc kubenswrapper[4818]: I1203 07:28:38.870558 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kh8ck\" (UniqueName: \"kubernetes.io/projected/a65c921d-45bd-4459-b994-98eb2cf2f9e4-kube-api-access-kh8ck\") on node \"crc\" DevicePath \"\"" Dec 03 07:28:39 crc kubenswrapper[4818]: I1203 07:28:39.516368 4818 scope.go:117] "RemoveContainer" containerID="3ad7210bf67e974d272a456badab288d16e6eb64d4ce8f685ab2c73b20f50f12" Dec 03 07:28:39 crc kubenswrapper[4818]: I1203 07:28:39.516405 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffrv8/crc-debug-vghqr" Dec 03 07:28:39 crc kubenswrapper[4818]: I1203 07:28:39.887969 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ffrv8/crc-debug-gb5bq"] Dec 03 07:28:39 crc kubenswrapper[4818]: E1203 07:28:39.888386 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a65c921d-45bd-4459-b994-98eb2cf2f9e4" containerName="container-00" Dec 03 07:28:39 crc kubenswrapper[4818]: I1203 07:28:39.888420 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="a65c921d-45bd-4459-b994-98eb2cf2f9e4" containerName="container-00" Dec 03 07:28:39 crc kubenswrapper[4818]: I1203 07:28:39.888625 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="a65c921d-45bd-4459-b994-98eb2cf2f9e4" containerName="container-00" Dec 03 07:28:39 crc kubenswrapper[4818]: I1203 07:28:39.889273 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffrv8/crc-debug-gb5bq" Dec 03 07:28:39 crc kubenswrapper[4818]: I1203 07:28:39.895075 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ffrv8"/"default-dockercfg-gw7ft" Dec 03 07:28:39 crc kubenswrapper[4818]: I1203 07:28:39.995526 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da52d1ae-8ad8-4829-8cb3-e6be4a9ef475-host\") pod \"crc-debug-gb5bq\" (UID: \"da52d1ae-8ad8-4829-8cb3-e6be4a9ef475\") " pod="openshift-must-gather-ffrv8/crc-debug-gb5bq" Dec 03 07:28:39 crc kubenswrapper[4818]: I1203 07:28:39.995655 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjqkb\" (UniqueName: \"kubernetes.io/projected/da52d1ae-8ad8-4829-8cb3-e6be4a9ef475-kube-api-access-zjqkb\") pod \"crc-debug-gb5bq\" (UID: \"da52d1ae-8ad8-4829-8cb3-e6be4a9ef475\") " pod="openshift-must-gather-ffrv8/crc-debug-gb5bq" Dec 03 07:28:40 crc kubenswrapper[4818]: I1203 07:28:40.097171 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da52d1ae-8ad8-4829-8cb3-e6be4a9ef475-host\") pod \"crc-debug-gb5bq\" (UID: \"da52d1ae-8ad8-4829-8cb3-e6be4a9ef475\") " pod="openshift-must-gather-ffrv8/crc-debug-gb5bq" Dec 03 07:28:40 crc kubenswrapper[4818]: I1203 07:28:40.097271 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjqkb\" (UniqueName: \"kubernetes.io/projected/da52d1ae-8ad8-4829-8cb3-e6be4a9ef475-kube-api-access-zjqkb\") pod \"crc-debug-gb5bq\" (UID: \"da52d1ae-8ad8-4829-8cb3-e6be4a9ef475\") " pod="openshift-must-gather-ffrv8/crc-debug-gb5bq" Dec 03 07:28:40 crc kubenswrapper[4818]: I1203 07:28:40.097371 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da52d1ae-8ad8-4829-8cb3-e6be4a9ef475-host\") pod \"crc-debug-gb5bq\" (UID: \"da52d1ae-8ad8-4829-8cb3-e6be4a9ef475\") " pod="openshift-must-gather-ffrv8/crc-debug-gb5bq" Dec 03 07:28:40 crc kubenswrapper[4818]: I1203 07:28:40.137538 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjqkb\" (UniqueName: \"kubernetes.io/projected/da52d1ae-8ad8-4829-8cb3-e6be4a9ef475-kube-api-access-zjqkb\") pod \"crc-debug-gb5bq\" (UID: \"da52d1ae-8ad8-4829-8cb3-e6be4a9ef475\") " pod="openshift-must-gather-ffrv8/crc-debug-gb5bq" Dec 03 07:28:40 crc kubenswrapper[4818]: I1203 07:28:40.205760 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffrv8/crc-debug-gb5bq" Dec 03 07:28:40 crc kubenswrapper[4818]: I1203 07:28:40.527718 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffrv8/crc-debug-gb5bq" event={"ID":"da52d1ae-8ad8-4829-8cb3-e6be4a9ef475","Type":"ContainerStarted","Data":"cf024dbe86517dc3b0aa8e88950fada6cb0aaea0fd8e15f90f7901063fb35c21"} Dec 03 07:28:40 crc kubenswrapper[4818]: I1203 07:28:40.527766 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffrv8/crc-debug-gb5bq" event={"ID":"da52d1ae-8ad8-4829-8cb3-e6be4a9ef475","Type":"ContainerStarted","Data":"29966ee018ef205ceacf5643ba162c46dbbfdf3c9210e145079748d118840cfc"} Dec 03 07:28:40 crc kubenswrapper[4818]: I1203 07:28:40.541960 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ffrv8/crc-debug-gb5bq" podStartSLOduration=1.541936542 podStartE2EDuration="1.541936542s" podCreationTimestamp="2025-12-03 07:28:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:28:40.539309736 +0000 UTC m=+3678.230918488" watchObservedRunningTime="2025-12-03 07:28:40.541936542 +0000 UTC m=+3678.233545304" Dec 03 07:28:40 crc kubenswrapper[4818]: I1203 07:28:40.755447 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a65c921d-45bd-4459-b994-98eb2cf2f9e4" path="/var/lib/kubelet/pods/a65c921d-45bd-4459-b994-98eb2cf2f9e4/volumes" Dec 03 07:28:41 crc kubenswrapper[4818]: I1203 07:28:41.540146 4818 generic.go:334] "Generic (PLEG): container finished" podID="da52d1ae-8ad8-4829-8cb3-e6be4a9ef475" containerID="cf024dbe86517dc3b0aa8e88950fada6cb0aaea0fd8e15f90f7901063fb35c21" exitCode=0 Dec 03 07:28:41 crc kubenswrapper[4818]: I1203 07:28:41.540189 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffrv8/crc-debug-gb5bq" event={"ID":"da52d1ae-8ad8-4829-8cb3-e6be4a9ef475","Type":"ContainerDied","Data":"cf024dbe86517dc3b0aa8e88950fada6cb0aaea0fd8e15f90f7901063fb35c21"} Dec 03 07:28:42 crc kubenswrapper[4818]: I1203 07:28:42.682633 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffrv8/crc-debug-gb5bq" Dec 03 07:28:42 crc kubenswrapper[4818]: I1203 07:28:42.717599 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ffrv8/crc-debug-gb5bq"] Dec 03 07:28:42 crc kubenswrapper[4818]: I1203 07:28:42.726169 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ffrv8/crc-debug-gb5bq"] Dec 03 07:28:42 crc kubenswrapper[4818]: I1203 07:28:42.798380 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da52d1ae-8ad8-4829-8cb3-e6be4a9ef475-host\") pod \"da52d1ae-8ad8-4829-8cb3-e6be4a9ef475\" (UID: \"da52d1ae-8ad8-4829-8cb3-e6be4a9ef475\") " Dec 03 07:28:42 crc kubenswrapper[4818]: I1203 07:28:42.798459 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da52d1ae-8ad8-4829-8cb3-e6be4a9ef475-host" (OuterVolumeSpecName: "host") pod "da52d1ae-8ad8-4829-8cb3-e6be4a9ef475" (UID: "da52d1ae-8ad8-4829-8cb3-e6be4a9ef475"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:28:42 crc kubenswrapper[4818]: I1203 07:28:42.798568 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjqkb\" (UniqueName: \"kubernetes.io/projected/da52d1ae-8ad8-4829-8cb3-e6be4a9ef475-kube-api-access-zjqkb\") pod \"da52d1ae-8ad8-4829-8cb3-e6be4a9ef475\" (UID: \"da52d1ae-8ad8-4829-8cb3-e6be4a9ef475\") " Dec 03 07:28:42 crc kubenswrapper[4818]: I1203 07:28:42.798962 4818 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/da52d1ae-8ad8-4829-8cb3-e6be4a9ef475-host\") on node \"crc\" DevicePath \"\"" Dec 03 07:28:42 crc kubenswrapper[4818]: I1203 07:28:42.813986 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da52d1ae-8ad8-4829-8cb3-e6be4a9ef475-kube-api-access-zjqkb" (OuterVolumeSpecName: "kube-api-access-zjqkb") pod "da52d1ae-8ad8-4829-8cb3-e6be4a9ef475" (UID: "da52d1ae-8ad8-4829-8cb3-e6be4a9ef475"). InnerVolumeSpecName "kube-api-access-zjqkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:28:42 crc kubenswrapper[4818]: I1203 07:28:42.901171 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjqkb\" (UniqueName: \"kubernetes.io/projected/da52d1ae-8ad8-4829-8cb3-e6be4a9ef475-kube-api-access-zjqkb\") on node \"crc\" DevicePath \"\"" Dec 03 07:28:43 crc kubenswrapper[4818]: I1203 07:28:43.562841 4818 scope.go:117] "RemoveContainer" containerID="cf024dbe86517dc3b0aa8e88950fada6cb0aaea0fd8e15f90f7901063fb35c21" Dec 03 07:28:43 crc kubenswrapper[4818]: I1203 07:28:43.562919 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffrv8/crc-debug-gb5bq" Dec 03 07:28:43 crc kubenswrapper[4818]: I1203 07:28:43.936708 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ffrv8/crc-debug-wk8wd"] Dec 03 07:28:43 crc kubenswrapper[4818]: E1203 07:28:43.937521 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da52d1ae-8ad8-4829-8cb3-e6be4a9ef475" containerName="container-00" Dec 03 07:28:43 crc kubenswrapper[4818]: I1203 07:28:43.937536 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="da52d1ae-8ad8-4829-8cb3-e6be4a9ef475" containerName="container-00" Dec 03 07:28:43 crc kubenswrapper[4818]: I1203 07:28:43.937698 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="da52d1ae-8ad8-4829-8cb3-e6be4a9ef475" containerName="container-00" Dec 03 07:28:43 crc kubenswrapper[4818]: I1203 07:28:43.938351 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffrv8/crc-debug-wk8wd" Dec 03 07:28:43 crc kubenswrapper[4818]: I1203 07:28:43.940303 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-ffrv8"/"default-dockercfg-gw7ft" Dec 03 07:28:44 crc kubenswrapper[4818]: I1203 07:28:44.021941 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13-host\") pod \"crc-debug-wk8wd\" (UID: \"1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13\") " pod="openshift-must-gather-ffrv8/crc-debug-wk8wd" Dec 03 07:28:44 crc kubenswrapper[4818]: I1203 07:28:44.022108 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cchrt\" (UniqueName: \"kubernetes.io/projected/1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13-kube-api-access-cchrt\") pod \"crc-debug-wk8wd\" (UID: \"1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13\") " pod="openshift-must-gather-ffrv8/crc-debug-wk8wd" Dec 03 07:28:44 crc kubenswrapper[4818]: I1203 07:28:44.123722 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13-host\") pod \"crc-debug-wk8wd\" (UID: \"1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13\") " pod="openshift-must-gather-ffrv8/crc-debug-wk8wd" Dec 03 07:28:44 crc kubenswrapper[4818]: I1203 07:28:44.123878 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13-host\") pod \"crc-debug-wk8wd\" (UID: \"1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13\") " pod="openshift-must-gather-ffrv8/crc-debug-wk8wd" Dec 03 07:28:44 crc kubenswrapper[4818]: I1203 07:28:44.123966 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cchrt\" (UniqueName: \"kubernetes.io/projected/1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13-kube-api-access-cchrt\") pod \"crc-debug-wk8wd\" (UID: \"1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13\") " pod="openshift-must-gather-ffrv8/crc-debug-wk8wd" Dec 03 07:28:44 crc kubenswrapper[4818]: I1203 07:28:44.145260 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cchrt\" (UniqueName: \"kubernetes.io/projected/1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13-kube-api-access-cchrt\") pod \"crc-debug-wk8wd\" (UID: \"1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13\") " pod="openshift-must-gather-ffrv8/crc-debug-wk8wd" Dec 03 07:28:44 crc kubenswrapper[4818]: I1203 07:28:44.258496 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffrv8/crc-debug-wk8wd" Dec 03 07:28:44 crc kubenswrapper[4818]: W1203 07:28:44.285628 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f3d2e72_6f7b_43a2_a7bc_b8e5e64f8f13.slice/crio-f3fa7cc112ef632abb0a2349f6b78b35a1e6324c161227fffa0748aaa66dde0f WatchSource:0}: Error finding container f3fa7cc112ef632abb0a2349f6b78b35a1e6324c161227fffa0748aaa66dde0f: Status 404 returned error can't find the container with id f3fa7cc112ef632abb0a2349f6b78b35a1e6324c161227fffa0748aaa66dde0f Dec 03 07:28:44 crc kubenswrapper[4818]: I1203 07:28:44.574415 4818 generic.go:334] "Generic (PLEG): container finished" podID="1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13" containerID="6ca968bd162914d097bb0004d03cf66e0558fff00a45a6b5a51253467dfdf040" exitCode=0 Dec 03 07:28:44 crc kubenswrapper[4818]: I1203 07:28:44.574492 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffrv8/crc-debug-wk8wd" event={"ID":"1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13","Type":"ContainerDied","Data":"6ca968bd162914d097bb0004d03cf66e0558fff00a45a6b5a51253467dfdf040"} Dec 03 07:28:44 crc kubenswrapper[4818]: I1203 07:28:44.574874 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffrv8/crc-debug-wk8wd" event={"ID":"1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13","Type":"ContainerStarted","Data":"f3fa7cc112ef632abb0a2349f6b78b35a1e6324c161227fffa0748aaa66dde0f"} Dec 03 07:28:44 crc kubenswrapper[4818]: I1203 07:28:44.616868 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ffrv8/crc-debug-wk8wd"] Dec 03 07:28:44 crc kubenswrapper[4818]: I1203 07:28:44.626558 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ffrv8/crc-debug-wk8wd"] Dec 03 07:28:44 crc kubenswrapper[4818]: I1203 07:28:44.752113 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da52d1ae-8ad8-4829-8cb3-e6be4a9ef475" path="/var/lib/kubelet/pods/da52d1ae-8ad8-4829-8cb3-e6be4a9ef475/volumes" Dec 03 07:28:45 crc kubenswrapper[4818]: I1203 07:28:45.697844 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffrv8/crc-debug-wk8wd" Dec 03 07:28:45 crc kubenswrapper[4818]: I1203 07:28:45.755058 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13-host\") pod \"1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13\" (UID: \"1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13\") " Dec 03 07:28:45 crc kubenswrapper[4818]: I1203 07:28:45.755195 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13-host" (OuterVolumeSpecName: "host") pod "1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13" (UID: "1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:28:45 crc kubenswrapper[4818]: I1203 07:28:45.755256 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cchrt\" (UniqueName: \"kubernetes.io/projected/1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13-kube-api-access-cchrt\") pod \"1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13\" (UID: \"1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13\") " Dec 03 07:28:45 crc kubenswrapper[4818]: I1203 07:28:45.756660 4818 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13-host\") on node \"crc\" DevicePath \"\"" Dec 03 07:28:45 crc kubenswrapper[4818]: I1203 07:28:45.761931 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13-kube-api-access-cchrt" (OuterVolumeSpecName: "kube-api-access-cchrt") pod "1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13" (UID: "1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13"). InnerVolumeSpecName "kube-api-access-cchrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:28:45 crc kubenswrapper[4818]: I1203 07:28:45.857304 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cchrt\" (UniqueName: \"kubernetes.io/projected/1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13-kube-api-access-cchrt\") on node \"crc\" DevicePath \"\"" Dec 03 07:28:46 crc kubenswrapper[4818]: I1203 07:28:46.602228 4818 scope.go:117] "RemoveContainer" containerID="6ca968bd162914d097bb0004d03cf66e0558fff00a45a6b5a51253467dfdf040" Dec 03 07:28:46 crc kubenswrapper[4818]: I1203 07:28:46.602260 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffrv8/crc-debug-wk8wd" Dec 03 07:28:46 crc kubenswrapper[4818]: I1203 07:28:46.757024 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13" path="/var/lib/kubelet/pods/1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13/volumes" Dec 03 07:28:47 crc kubenswrapper[4818]: I1203 07:28:47.737986 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:28:47 crc kubenswrapper[4818]: E1203 07:28:47.738297 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:29:00 crc kubenswrapper[4818]: I1203 07:29:00.170388 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7454b897fd-qzp7b_5e283b07-3849-4796-9b82-84e5ca0bf431/barbican-api/0.log" Dec 03 07:29:00 crc kubenswrapper[4818]: I1203 07:29:00.238631 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7454b897fd-qzp7b_5e283b07-3849-4796-9b82-84e5ca0bf431/barbican-api-log/0.log" Dec 03 07:29:00 crc kubenswrapper[4818]: I1203 07:29:00.351424 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7b4848f678-kfvpp_6d8d8f21-1f1f-4431-8079-ffad453ff3f1/barbican-keystone-listener/0.log" Dec 03 07:29:00 crc kubenswrapper[4818]: I1203 07:29:00.380079 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7b4848f678-kfvpp_6d8d8f21-1f1f-4431-8079-ffad453ff3f1/barbican-keystone-listener-log/0.log" Dec 03 07:29:00 crc kubenswrapper[4818]: I1203 07:29:00.549845 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-554df66f87-jscgb_d8e58804-c7d8-48ce-9db3-6c3ab56f1c13/barbican-worker/0.log" Dec 03 07:29:00 crc kubenswrapper[4818]: I1203 07:29:00.551069 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-554df66f87-jscgb_d8e58804-c7d8-48ce-9db3-6c3ab56f1c13/barbican-worker-log/0.log" Dec 03 07:29:00 crc kubenswrapper[4818]: I1203 07:29:00.707941 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g_2b2dc637-0ece-412a-b1e2-413dff6a4f08/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:29:00 crc kubenswrapper[4818]: I1203 07:29:00.794910 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_631b9437-fe40-43b7-b8ad-34b4cad6ffa1/ceilometer-central-agent/0.log" Dec 03 07:29:00 crc kubenswrapper[4818]: I1203 07:29:00.875596 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_631b9437-fe40-43b7-b8ad-34b4cad6ffa1/ceilometer-notification-agent/0.log" Dec 03 07:29:00 crc kubenswrapper[4818]: I1203 07:29:00.921478 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_631b9437-fe40-43b7-b8ad-34b4cad6ffa1/proxy-httpd/0.log" Dec 03 07:29:00 crc kubenswrapper[4818]: I1203 07:29:00.984342 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_631b9437-fe40-43b7-b8ad-34b4cad6ffa1/sg-core/0.log" Dec 03 07:29:01 crc kubenswrapper[4818]: I1203 07:29:01.145867 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_462028f4-61c0-4862-8707-3981513c3b76/cinder-api-log/0.log" Dec 03 07:29:01 crc kubenswrapper[4818]: I1203 07:29:01.155017 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_462028f4-61c0-4862-8707-3981513c3b76/cinder-api/0.log" Dec 03 07:29:01 crc kubenswrapper[4818]: I1203 07:29:01.336480 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5ad1caca-cc5b-4625-8776-2fdc829594a4/cinder-scheduler/0.log" Dec 03 07:29:01 crc kubenswrapper[4818]: I1203 07:29:01.385034 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5ad1caca-cc5b-4625-8776-2fdc829594a4/probe/0.log" Dec 03 07:29:01 crc kubenswrapper[4818]: I1203 07:29:01.466226 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn_71c50082-f6aa-4099-be63-2b33234f0636/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:29:01 crc kubenswrapper[4818]: I1203 07:29:01.623397 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9_2ef6c961-25b1-4f6f-a98b-65d79787e9e1/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:29:01 crc kubenswrapper[4818]: I1203 07:29:01.708351 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-gbkbl_19b91d65-7f62-47e7-b12c-33135811b76b/init/0.log" Dec 03 07:29:01 crc kubenswrapper[4818]: I1203 07:29:01.911106 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-gbkbl_19b91d65-7f62-47e7-b12c-33135811b76b/init/0.log" Dec 03 07:29:01 crc kubenswrapper[4818]: I1203 07:29:01.951283 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq_86dc24e9-831f-4731-a612-98009562639c/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:29:02 crc kubenswrapper[4818]: I1203 07:29:02.016302 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-gbkbl_19b91d65-7f62-47e7-b12c-33135811b76b/dnsmasq-dns/0.log" Dec 03 07:29:02 crc kubenswrapper[4818]: I1203 07:29:02.130403 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0c46a9a7-6f86-40b8-8957-5d0f11d11262/glance-httpd/0.log" Dec 03 07:29:02 crc kubenswrapper[4818]: I1203 07:29:02.193404 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0c46a9a7-6f86-40b8-8957-5d0f11d11262/glance-log/0.log" Dec 03 07:29:02 crc kubenswrapper[4818]: I1203 07:29:02.334493 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_22a6b677-2df1-41ed-a420-dbc5ce60afc5/glance-httpd/0.log" Dec 03 07:29:02 crc kubenswrapper[4818]: I1203 07:29:02.351604 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_22a6b677-2df1-41ed-a420-dbc5ce60afc5/glance-log/0.log" Dec 03 07:29:02 crc kubenswrapper[4818]: I1203 07:29:02.585121 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-68b68c8c84-qt25s_23c89695-46b0-49d8-ab09-eb002d3cb94e/horizon/0.log" Dec 03 07:29:02 crc kubenswrapper[4818]: I1203 07:29:02.669405 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-z89w5_1808af84-64d0-4b2a-a5bf-32653bb3c948/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:29:02 crc kubenswrapper[4818]: I1203 07:29:02.757376 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:29:02 crc kubenswrapper[4818]: E1203 07:29:02.757943 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:29:02 crc kubenswrapper[4818]: I1203 07:29:02.881756 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-pdfz4_9568e2c6-4f91-49ce-80e6-0081b7bef2c5/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:29:02 crc kubenswrapper[4818]: I1203 07:29:02.909027 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-68b68c8c84-qt25s_23c89695-46b0-49d8-ab09-eb002d3cb94e/horizon-log/0.log" Dec 03 07:29:03 crc kubenswrapper[4818]: I1203 07:29:03.163099 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29412421-4n5xh_658c06c1-51a3-467d-8de0-2205f7c22cbc/keystone-cron/0.log" Dec 03 07:29:03 crc kubenswrapper[4818]: I1203 07:29:03.180443 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-75b5b56458-pbq6h_29098cb1-d622-4ee2-ad26-f8d51eec41f4/keystone-api/0.log" Dec 03 07:29:03 crc kubenswrapper[4818]: I1203 07:29:03.469902 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_0d105427-cef6-4ec6-8997-08f8a94e5b56/kube-state-metrics/0.log" Dec 03 07:29:03 crc kubenswrapper[4818]: I1203 07:29:03.564427 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd_e5796328-5b74-4f5a-939b-403485be16a2/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:29:03 crc kubenswrapper[4818]: I1203 07:29:03.922284 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-77cd7f4557-lnszf_bcf80b5c-a427-4bcb-a60f-51b730b41cf2/neutron-httpd/0.log" Dec 03 07:29:03 crc kubenswrapper[4818]: I1203 07:29:03.953949 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-77cd7f4557-lnszf_bcf80b5c-a427-4bcb-a60f-51b730b41cf2/neutron-api/0.log" Dec 03 07:29:04 crc kubenswrapper[4818]: I1203 07:29:04.074952 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9_31a69e56-fab5-4dd2-83bb-78cce024c2c1/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:29:04 crc kubenswrapper[4818]: I1203 07:29:04.505397 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_5a5a4f96-9611-456a-a084-88ba9f3b548a/nova-api-log/0.log" Dec 03 07:29:04 crc kubenswrapper[4818]: I1203 07:29:04.594504 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_f40ddc54-764b-4718-aea1-ac8e81eccae7/nova-cell0-conductor-conductor/0.log" Dec 03 07:29:04 crc kubenswrapper[4818]: I1203 07:29:04.767341 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_5a5a4f96-9611-456a-a084-88ba9f3b548a/nova-api-api/0.log" Dec 03 07:29:04 crc kubenswrapper[4818]: I1203 07:29:04.798641 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_d772d6d8-be96-4256-84a5-e9f4713ad4ec/nova-cell1-conductor-conductor/0.log" Dec 03 07:29:04 crc kubenswrapper[4818]: I1203 07:29:04.898148 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_b1f0f202-1c95-48ed-96fa-1606b2c81a6f/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 07:29:05 crc kubenswrapper[4818]: I1203 07:29:05.055604 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-l4l2p_febe4f5e-461e-4562-a168-b99a22ef6e7c/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:29:05 crc kubenswrapper[4818]: I1203 07:29:05.206020 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_78622406-d448-4cf4-b5c9-c639b0fb6a41/nova-metadata-log/0.log" Dec 03 07:29:05 crc kubenswrapper[4818]: I1203 07:29:05.541857 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_39a9a10c-e02f-4468-95a9-651b694b1f9a/nova-scheduler-scheduler/0.log" Dec 03 07:29:05 crc kubenswrapper[4818]: I1203 07:29:05.592340 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a0dd3e13-e5ee-4315-ab7a-222c902192be/mysql-bootstrap/0.log" Dec 03 07:29:05 crc kubenswrapper[4818]: I1203 07:29:05.753265 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a0dd3e13-e5ee-4315-ab7a-222c902192be/mysql-bootstrap/0.log" Dec 03 07:29:05 crc kubenswrapper[4818]: I1203 07:29:05.783482 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a0dd3e13-e5ee-4315-ab7a-222c902192be/galera/0.log" Dec 03 07:29:05 crc kubenswrapper[4818]: I1203 07:29:05.995270 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_206b044a-6264-4c4e-89b9-7cacc4f30ac5/mysql-bootstrap/0.log" Dec 03 07:29:06 crc kubenswrapper[4818]: I1203 07:29:06.136043 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_206b044a-6264-4c4e-89b9-7cacc4f30ac5/mysql-bootstrap/0.log" Dec 03 07:29:06 crc kubenswrapper[4818]: I1203 07:29:06.136120 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_206b044a-6264-4c4e-89b9-7cacc4f30ac5/galera/0.log" Dec 03 07:29:06 crc kubenswrapper[4818]: I1203 07:29:06.355790 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_aecf49dc-5a72-41aa-bc7c-c42bf226ea71/openstackclient/0.log" Dec 03 07:29:06 crc kubenswrapper[4818]: I1203 07:29:06.466803 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_78622406-d448-4cf4-b5c9-c639b0fb6a41/nova-metadata-metadata/0.log" Dec 03 07:29:06 crc kubenswrapper[4818]: I1203 07:29:06.670309 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-kvg5t_04bf523a-de6a-49a0-93ad-cca1e22939c6/ovn-controller/0.log" Dec 03 07:29:06 crc kubenswrapper[4818]: I1203 07:29:06.813029 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-srrh4_997e3f50-bc25-4436-b03d-5e98770df25f/openstack-network-exporter/0.log" Dec 03 07:29:06 crc kubenswrapper[4818]: I1203 07:29:06.932664 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-brxh4_b5f31813-ba17-4452-9b1a-ee7dcfac564e/ovsdb-server-init/0.log" Dec 03 07:29:07 crc kubenswrapper[4818]: I1203 07:29:07.120163 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-brxh4_b5f31813-ba17-4452-9b1a-ee7dcfac564e/ovs-vswitchd/0.log" Dec 03 07:29:07 crc kubenswrapper[4818]: I1203 07:29:07.167335 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-brxh4_b5f31813-ba17-4452-9b1a-ee7dcfac564e/ovsdb-server/0.log" Dec 03 07:29:07 crc kubenswrapper[4818]: I1203 07:29:07.167904 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-brxh4_b5f31813-ba17-4452-9b1a-ee7dcfac564e/ovsdb-server-init/0.log" Dec 03 07:29:07 crc kubenswrapper[4818]: I1203 07:29:07.420936 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8df18ace-e467-4e73-9304-3ea16df510e8/openstack-network-exporter/0.log" Dec 03 07:29:07 crc kubenswrapper[4818]: I1203 07:29:07.422058 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8df18ace-e467-4e73-9304-3ea16df510e8/ovn-northd/0.log" Dec 03 07:29:07 crc kubenswrapper[4818]: I1203 07:29:07.434009 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-xxxp7_22cbc8e9-1434-4dcb-8fd3-24def21fe451/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:29:07 crc kubenswrapper[4818]: I1203 07:29:07.585093 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9b27dc29-202f-4efb-9545-031eaa1d7da3/openstack-network-exporter/0.log" Dec 03 07:29:07 crc kubenswrapper[4818]: I1203 07:29:07.692313 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9b27dc29-202f-4efb-9545-031eaa1d7da3/ovsdbserver-nb/0.log" Dec 03 07:29:07 crc kubenswrapper[4818]: I1203 07:29:07.757245 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_a27ef44f-2bff-4714-8018-a50f2f2de24a/openstack-network-exporter/0.log" Dec 03 07:29:07 crc kubenswrapper[4818]: I1203 07:29:07.869985 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_a27ef44f-2bff-4714-8018-a50f2f2de24a/ovsdbserver-sb/0.log" Dec 03 07:29:08 crc kubenswrapper[4818]: I1203 07:29:08.014640 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-748bc8fb48-scr66_e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff/placement-log/0.log" Dec 03 07:29:08 crc kubenswrapper[4818]: I1203 07:29:08.120798 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-748bc8fb48-scr66_e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff/placement-api/0.log" Dec 03 07:29:08 crc kubenswrapper[4818]: I1203 07:29:08.199301 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8078039a-8674-40c1-ba1a-ff8244d17e16/setup-container/0.log" Dec 03 07:29:08 crc kubenswrapper[4818]: I1203 07:29:08.491404 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8078039a-8674-40c1-ba1a-ff8244d17e16/rabbitmq/0.log" Dec 03 07:29:08 crc kubenswrapper[4818]: I1203 07:29:08.498700 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8078039a-8674-40c1-ba1a-ff8244d17e16/setup-container/0.log" Dec 03 07:29:08 crc kubenswrapper[4818]: I1203 07:29:08.531825 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e6fa481e-32ce-4840-b457-0a692558cdb5/setup-container/0.log" Dec 03 07:29:08 crc kubenswrapper[4818]: I1203 07:29:08.738994 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e6fa481e-32ce-4840-b457-0a692558cdb5/setup-container/0.log" Dec 03 07:29:08 crc kubenswrapper[4818]: I1203 07:29:08.814030 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e6fa481e-32ce-4840-b457-0a692558cdb5/rabbitmq/0.log" Dec 03 07:29:08 crc kubenswrapper[4818]: I1203 07:29:08.839138 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26_3bf68a0c-4bbc-4f9e-b3be-503d857ead66/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:29:09 crc kubenswrapper[4818]: I1203 07:29:09.082378 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-nsd6q_fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:29:09 crc kubenswrapper[4818]: I1203 07:29:09.083099 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh_37db016a-a063-4751-bf39-aadd6746fa13/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:29:09 crc kubenswrapper[4818]: I1203 07:29:09.304913 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-d689l_90add555-879b-4da5-a350-40ce938e9956/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:29:09 crc kubenswrapper[4818]: I1203 07:29:09.361973 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-lbkfx_b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de/ssh-known-hosts-edpm-deployment/0.log" Dec 03 07:29:09 crc kubenswrapper[4818]: I1203 07:29:09.615622 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7954d7ccff-th9lw_0ba50d38-959d-49f4-a98a-6280a78661e4/proxy-server/0.log" Dec 03 07:29:09 crc kubenswrapper[4818]: I1203 07:29:09.679503 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7954d7ccff-th9lw_0ba50d38-959d-49f4-a98a-6280a78661e4/proxy-httpd/0.log" Dec 03 07:29:09 crc kubenswrapper[4818]: I1203 07:29:09.764541 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-pkxxx_337c7d6b-71fc-4044-98ec-de8bfbf5dd6f/swift-ring-rebalance/0.log" Dec 03 07:29:09 crc kubenswrapper[4818]: I1203 07:29:09.856655 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/account-auditor/0.log" Dec 03 07:29:09 crc kubenswrapper[4818]: I1203 07:29:09.893924 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/account-reaper/0.log" Dec 03 07:29:10 crc kubenswrapper[4818]: I1203 07:29:10.172637 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/account-replicator/0.log" Dec 03 07:29:10 crc kubenswrapper[4818]: I1203 07:29:10.217279 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/container-auditor/0.log" Dec 03 07:29:10 crc kubenswrapper[4818]: I1203 07:29:10.285161 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/account-server/0.log" Dec 03 07:29:10 crc kubenswrapper[4818]: I1203 07:29:10.326977 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/container-replicator/0.log" Dec 03 07:29:10 crc kubenswrapper[4818]: I1203 07:29:10.389271 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/container-server/0.log" Dec 03 07:29:10 crc kubenswrapper[4818]: I1203 07:29:10.422574 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/container-updater/0.log" Dec 03 07:29:10 crc kubenswrapper[4818]: I1203 07:29:10.727675 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/object-expirer/0.log" Dec 03 07:29:10 crc kubenswrapper[4818]: I1203 07:29:10.730978 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/object-server/0.log" Dec 03 07:29:10 crc kubenswrapper[4818]: I1203 07:29:10.788713 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/object-replicator/0.log" Dec 03 07:29:10 crc kubenswrapper[4818]: I1203 07:29:10.823463 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/object-auditor/0.log" Dec 03 07:29:10 crc kubenswrapper[4818]: I1203 07:29:10.934552 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/rsync/0.log" Dec 03 07:29:10 crc kubenswrapper[4818]: I1203 07:29:10.965781 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/object-updater/0.log" Dec 03 07:29:10 crc kubenswrapper[4818]: I1203 07:29:10.986997 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/swift-recon-cron/0.log" Dec 03 07:29:11 crc kubenswrapper[4818]: I1203 07:29:11.250451 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-wkr94_9113c780-f00c-4f24-b000-26270c809028/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:29:11 crc kubenswrapper[4818]: I1203 07:29:11.254572 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf/tempest-tests-tempest-tests-runner/0.log" Dec 03 07:29:11 crc kubenswrapper[4818]: I1203 07:29:11.477304 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_78a39ae7-b979-4d32-a128-25dca0a393e4/test-operator-logs-container/0.log" Dec 03 07:29:11 crc kubenswrapper[4818]: I1203 07:29:11.508899 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh_4e2bf3bf-bd96-4042-b1e6-11247aefdbe2/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:29:15 crc kubenswrapper[4818]: I1203 07:29:15.737180 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:29:15 crc kubenswrapper[4818]: E1203 07:29:15.738039 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:29:19 crc kubenswrapper[4818]: I1203 07:29:19.266608 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_c826858d-45cb-4ae3-8fa2-1531f9390480/memcached/0.log" Dec 03 07:29:30 crc kubenswrapper[4818]: I1203 07:29:30.738265 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:29:30 crc kubenswrapper[4818]: E1203 07:29:30.739017 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:29:36 crc kubenswrapper[4818]: I1203 07:29:36.807757 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-z6c4f_7c36b14e-4e1d-4956-a505-ecdffb9d9f9c/kube-rbac-proxy/0.log" Dec 03 07:29:36 crc kubenswrapper[4818]: I1203 07:29:36.895444 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-z6c4f_7c36b14e-4e1d-4956-a505-ecdffb9d9f9c/manager/0.log" Dec 03 07:29:37 crc kubenswrapper[4818]: I1203 07:29:37.006000 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-4hdr6_4130e7e0-4596-451f-b346-5afba08e9dea/kube-rbac-proxy/0.log" Dec 03 07:29:37 crc kubenswrapper[4818]: I1203 07:29:37.039264 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-4hdr6_4130e7e0-4596-451f-b346-5afba08e9dea/manager/0.log" Dec 03 07:29:37 crc kubenswrapper[4818]: I1203 07:29:37.189850 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk_413ec55a-90aa-4e14-a447-47595f913b9a/util/0.log" Dec 03 07:29:37 crc kubenswrapper[4818]: I1203 07:29:37.327532 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk_413ec55a-90aa-4e14-a447-47595f913b9a/util/0.log" Dec 03 07:29:37 crc kubenswrapper[4818]: I1203 07:29:37.328336 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk_413ec55a-90aa-4e14-a447-47595f913b9a/pull/0.log" Dec 03 07:29:37 crc kubenswrapper[4818]: I1203 07:29:37.328396 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk_413ec55a-90aa-4e14-a447-47595f913b9a/pull/0.log" Dec 03 07:29:37 crc kubenswrapper[4818]: I1203 07:29:37.506967 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk_413ec55a-90aa-4e14-a447-47595f913b9a/util/0.log" Dec 03 07:29:37 crc kubenswrapper[4818]: I1203 07:29:37.523247 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk_413ec55a-90aa-4e14-a447-47595f913b9a/extract/0.log" Dec 03 07:29:37 crc kubenswrapper[4818]: I1203 07:29:37.551414 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk_413ec55a-90aa-4e14-a447-47595f913b9a/pull/0.log" Dec 03 07:29:37 crc kubenswrapper[4818]: I1203 07:29:37.667550 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-wndt9_91e645ae-28ec-4d3b-8fe3-e75aa1acf5a5/kube-rbac-proxy/0.log" Dec 03 07:29:37 crc kubenswrapper[4818]: I1203 07:29:37.700185 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-wndt9_91e645ae-28ec-4d3b-8fe3-e75aa1acf5a5/manager/0.log" Dec 03 07:29:37 crc kubenswrapper[4818]: I1203 07:29:37.745663 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-2qvsd_528c7c89-7980-4acc-9321-ff56c9d38489/kube-rbac-proxy/0.log" Dec 03 07:29:37 crc kubenswrapper[4818]: I1203 07:29:37.922233 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-gtmmz_b29acffb-a7e4-4029-93fa-20f20de27dfe/manager/0.log" Dec 03 07:29:37 crc kubenswrapper[4818]: I1203 07:29:37.929433 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-gtmmz_b29acffb-a7e4-4029-93fa-20f20de27dfe/kube-rbac-proxy/0.log" Dec 03 07:29:37 crc kubenswrapper[4818]: I1203 07:29:37.976373 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-2qvsd_528c7c89-7980-4acc-9321-ff56c9d38489/manager/0.log" Dec 03 07:29:38 crc kubenswrapper[4818]: I1203 07:29:38.084398 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-thzjt_6d5bf0d5-df89-4ab4-a711-e6f7f3d1123e/kube-rbac-proxy/0.log" Dec 03 07:29:38 crc kubenswrapper[4818]: I1203 07:29:38.141645 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-thzjt_6d5bf0d5-df89-4ab4-a711-e6f7f3d1123e/manager/0.log" Dec 03 07:29:38 crc kubenswrapper[4818]: I1203 07:29:38.276251 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-gmkzd_a0b8fddb-4636-4f13-8d0b-3e3220721eca/kube-rbac-proxy/0.log" Dec 03 07:29:38 crc kubenswrapper[4818]: I1203 07:29:38.377507 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-hpnhz_3d1565f0-62d4-4ecd-9d73-16338035196e/kube-rbac-proxy/0.log" Dec 03 07:29:38 crc kubenswrapper[4818]: I1203 07:29:38.398478 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-gmkzd_a0b8fddb-4636-4f13-8d0b-3e3220721eca/manager/0.log" Dec 03 07:29:38 crc kubenswrapper[4818]: I1203 07:29:38.470112 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-hpnhz_3d1565f0-62d4-4ecd-9d73-16338035196e/manager/0.log" Dec 03 07:29:38 crc kubenswrapper[4818]: I1203 07:29:38.591600 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-27nrr_1bd3fb88-7598-4a28-90fe-324d855dc90e/kube-rbac-proxy/0.log" Dec 03 07:29:38 crc kubenswrapper[4818]: I1203 07:29:38.620675 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-27nrr_1bd3fb88-7598-4a28-90fe-324d855dc90e/manager/0.log" Dec 03 07:29:38 crc kubenswrapper[4818]: I1203 07:29:38.761657 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-s84mt_0c360185-7fff-4054-a85e-ca2bbb1fa5e0/kube-rbac-proxy/0.log" Dec 03 07:29:38 crc kubenswrapper[4818]: I1203 07:29:38.845930 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-s84mt_0c360185-7fff-4054-a85e-ca2bbb1fa5e0/manager/0.log" Dec 03 07:29:38 crc kubenswrapper[4818]: I1203 07:29:38.879646 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-98z98_df292cb6-9f13-4cec-a7e1-1ef3155a7b1f/kube-rbac-proxy/0.log" Dec 03 07:29:38 crc kubenswrapper[4818]: I1203 07:29:38.958444 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-98z98_df292cb6-9f13-4cec-a7e1-1ef3155a7b1f/manager/0.log" Dec 03 07:29:39 crc kubenswrapper[4818]: I1203 07:29:39.022735 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-b4thx_0d4b3be5-bb2e-4580-8142-60d15b16c36f/kube-rbac-proxy/0.log" Dec 03 07:29:39 crc kubenswrapper[4818]: I1203 07:29:39.091191 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-b4thx_0d4b3be5-bb2e-4580-8142-60d15b16c36f/manager/0.log" Dec 03 07:29:39 crc kubenswrapper[4818]: I1203 07:29:39.214193 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-xtq4s_1dc6c089-07b6-4818-a138-6b6a8c0cd1b5/kube-rbac-proxy/0.log" Dec 03 07:29:39 crc kubenswrapper[4818]: I1203 07:29:39.298599 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-xtq4s_1dc6c089-07b6-4818-a138-6b6a8c0cd1b5/manager/0.log" Dec 03 07:29:39 crc kubenswrapper[4818]: I1203 07:29:39.337979 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-6rr64_45dcf421-464d-4eb9-a23f-9924ecf5bb03/kube-rbac-proxy/0.log" Dec 03 07:29:39 crc kubenswrapper[4818]: I1203 07:29:39.403784 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-6rr64_45dcf421-464d-4eb9-a23f-9924ecf5bb03/manager/0.log" Dec 03 07:29:39 crc kubenswrapper[4818]: I1203 07:29:39.482739 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd445pgf_f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34/kube-rbac-proxy/0.log" Dec 03 07:29:39 crc kubenswrapper[4818]: I1203 07:29:39.510352 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd445pgf_f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34/manager/0.log" Dec 03 07:29:39 crc kubenswrapper[4818]: I1203 07:29:39.908907 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-555856bbb4-tlxbc_69cb7eae-cb67-42ef-b36d-d6287f1f14d4/operator/0.log" Dec 03 07:29:39 crc kubenswrapper[4818]: I1203 07:29:39.932131 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-jr8dk_f28e0eba-98e8-440c-8c04-91ce47ea304b/registry-server/0.log" Dec 03 07:29:40 crc kubenswrapper[4818]: I1203 07:29:40.131254 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-nchr7_4b33f4e2-edb8-4e90-b73a-4dc2b13edd41/kube-rbac-proxy/0.log" Dec 03 07:29:40 crc kubenswrapper[4818]: I1203 07:29:40.286004 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-nchr7_4b33f4e2-edb8-4e90-b73a-4dc2b13edd41/manager/0.log" Dec 03 07:29:40 crc kubenswrapper[4818]: I1203 07:29:40.352808 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-2d5xh_e977170c-c8e9-4888-951b-bffa4d4d21d7/kube-rbac-proxy/0.log" Dec 03 07:29:40 crc kubenswrapper[4818]: I1203 07:29:40.603842 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-2d5xh_e977170c-c8e9-4888-951b-bffa4d4d21d7/manager/0.log" Dec 03 07:29:40 crc kubenswrapper[4818]: I1203 07:29:40.687540 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-5q95v_5a1859a2-fa2e-4d8c-bd64-7b63297f258c/operator/0.log" Dec 03 07:29:40 crc kubenswrapper[4818]: I1203 07:29:40.820406 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-r4b9n_2e8a3376-fa5d-458d-a659-0c3cef8af10b/kube-rbac-proxy/0.log" Dec 03 07:29:40 crc kubenswrapper[4818]: I1203 07:29:40.837651 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-r4b9n_2e8a3376-fa5d-458d-a659-0c3cef8af10b/manager/0.log" Dec 03 07:29:40 crc kubenswrapper[4818]: I1203 07:29:40.929427 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-98d87bf44-8h8d6_3a635982-10c6-47cc-8462-74873b98410a/manager/0.log" Dec 03 07:29:41 crc kubenswrapper[4818]: I1203 07:29:41.039620 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-bjq7j_b78bb8f0-22f9-4b8e-93e2-23422db45354/manager/0.log" Dec 03 07:29:41 crc kubenswrapper[4818]: I1203 07:29:41.082398 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-bjq7j_b78bb8f0-22f9-4b8e-93e2-23422db45354/kube-rbac-proxy/0.log" Dec 03 07:29:41 crc kubenswrapper[4818]: I1203 07:29:41.158051 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-8hntr_84dd0135-3ec6-4418-98f9-de8474662893/kube-rbac-proxy/0.log" Dec 03 07:29:41 crc kubenswrapper[4818]: I1203 07:29:41.189680 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-8hntr_84dd0135-3ec6-4418-98f9-de8474662893/manager/0.log" Dec 03 07:29:41 crc kubenswrapper[4818]: I1203 07:29:41.266616 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-jd5hg_56bb5f70-9dd8-433f-a974-0c95065c1dd1/kube-rbac-proxy/0.log" Dec 03 07:29:41 crc kubenswrapper[4818]: I1203 07:29:41.311361 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-jd5hg_56bb5f70-9dd8-433f-a974-0c95065c1dd1/manager/0.log" Dec 03 07:29:44 crc kubenswrapper[4818]: I1203 07:29:44.737524 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:29:44 crc kubenswrapper[4818]: E1203 07:29:44.738333 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:29:58 crc kubenswrapper[4818]: I1203 07:29:58.736937 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:29:58 crc kubenswrapper[4818]: E1203 07:29:58.737736 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.169231 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm"] Dec 03 07:30:00 crc kubenswrapper[4818]: E1203 07:30:00.170044 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13" containerName="container-00" Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.170062 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13" containerName="container-00" Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.170332 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f3d2e72-6f7b-43a2-a7bc-b8e5e64f8f13" containerName="container-00" Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.171148 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm" Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.172947 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.173387 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.180843 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm"] Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.366878 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-secret-volume\") pod \"collect-profiles-29412450-r76tm\" (UID: \"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm" Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.367240 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rh6z\" (UniqueName: \"kubernetes.io/projected/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-kube-api-access-4rh6z\") pod \"collect-profiles-29412450-r76tm\" (UID: \"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm" Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.367375 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-config-volume\") pod \"collect-profiles-29412450-r76tm\" (UID: \"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm" Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.469576 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rh6z\" (UniqueName: \"kubernetes.io/projected/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-kube-api-access-4rh6z\") pod \"collect-profiles-29412450-r76tm\" (UID: \"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm" Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.469733 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-config-volume\") pod \"collect-profiles-29412450-r76tm\" (UID: \"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm" Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.469782 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-secret-volume\") pod \"collect-profiles-29412450-r76tm\" (UID: \"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm" Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.470770 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-config-volume\") pod \"collect-profiles-29412450-r76tm\" (UID: \"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm" Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.480562 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-secret-volume\") pod \"collect-profiles-29412450-r76tm\" (UID: \"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm" Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.494687 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rh6z\" (UniqueName: \"kubernetes.io/projected/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-kube-api-access-4rh6z\") pod \"collect-profiles-29412450-r76tm\" (UID: \"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm" Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.506196 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm" Dec 03 07:30:00 crc kubenswrapper[4818]: I1203 07:30:00.997263 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm"] Dec 03 07:30:01 crc kubenswrapper[4818]: I1203 07:30:01.296384 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm" event={"ID":"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2","Type":"ContainerStarted","Data":"2234a5bc547b1f75c9488beb0cf41b1cbe6093784e6345ab144febbc69e54f89"} Dec 03 07:30:01 crc kubenswrapper[4818]: I1203 07:30:01.296640 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm" event={"ID":"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2","Type":"ContainerStarted","Data":"578f35d8e0271bbd75028adc7c24d086b0f90dc84e95b7ca1d62dfea16e53168"} Dec 03 07:30:02 crc kubenswrapper[4818]: I1203 07:30:02.308386 4818 generic.go:334] "Generic (PLEG): container finished" podID="b9c72d04-9a02-4a2d-b458-9bba7d48d6a2" containerID="2234a5bc547b1f75c9488beb0cf41b1cbe6093784e6345ab144febbc69e54f89" exitCode=0 Dec 03 07:30:02 crc kubenswrapper[4818]: I1203 07:30:02.308460 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm" event={"ID":"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2","Type":"ContainerDied","Data":"2234a5bc547b1f75c9488beb0cf41b1cbe6093784e6345ab144febbc69e54f89"} Dec 03 07:30:03 crc kubenswrapper[4818]: I1203 07:30:03.263621 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-ltb69_faad9125-f494-49ae-ba77-941e6abe67b9/control-plane-machine-set-operator/0.log" Dec 03 07:30:03 crc kubenswrapper[4818]: I1203 07:30:03.494721 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jf227_8ec6e899-8380-4812-8f65-6fd72db12939/kube-rbac-proxy/0.log" Dec 03 07:30:03 crc kubenswrapper[4818]: I1203 07:30:03.515314 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jf227_8ec6e899-8380-4812-8f65-6fd72db12939/machine-api-operator/0.log" Dec 03 07:30:03 crc kubenswrapper[4818]: I1203 07:30:03.656150 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm" Dec 03 07:30:03 crc kubenswrapper[4818]: I1203 07:30:03.730225 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-secret-volume\") pod \"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2\" (UID: \"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2\") " Dec 03 07:30:03 crc kubenswrapper[4818]: I1203 07:30:03.730365 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rh6z\" (UniqueName: \"kubernetes.io/projected/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-kube-api-access-4rh6z\") pod \"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2\" (UID: \"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2\") " Dec 03 07:30:03 crc kubenswrapper[4818]: I1203 07:30:03.730460 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-config-volume\") pod \"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2\" (UID: \"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2\") " Dec 03 07:30:03 crc kubenswrapper[4818]: I1203 07:30:03.731361 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-config-volume" (OuterVolumeSpecName: "config-volume") pod "b9c72d04-9a02-4a2d-b458-9bba7d48d6a2" (UID: "b9c72d04-9a02-4a2d-b458-9bba7d48d6a2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 07:30:03 crc kubenswrapper[4818]: I1203 07:30:03.738140 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-kube-api-access-4rh6z" (OuterVolumeSpecName: "kube-api-access-4rh6z") pod "b9c72d04-9a02-4a2d-b458-9bba7d48d6a2" (UID: "b9c72d04-9a02-4a2d-b458-9bba7d48d6a2"). InnerVolumeSpecName "kube-api-access-4rh6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:30:03 crc kubenswrapper[4818]: I1203 07:30:03.739358 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b9c72d04-9a02-4a2d-b458-9bba7d48d6a2" (UID: "b9c72d04-9a02-4a2d-b458-9bba7d48d6a2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 07:30:03 crc kubenswrapper[4818]: I1203 07:30:03.832139 4818 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 07:30:03 crc kubenswrapper[4818]: I1203 07:30:03.832173 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rh6z\" (UniqueName: \"kubernetes.io/projected/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-kube-api-access-4rh6z\") on node \"crc\" DevicePath \"\"" Dec 03 07:30:03 crc kubenswrapper[4818]: I1203 07:30:03.832182 4818 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9c72d04-9a02-4a2d-b458-9bba7d48d6a2-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 07:30:04 crc kubenswrapper[4818]: I1203 07:30:04.326914 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm" event={"ID":"b9c72d04-9a02-4a2d-b458-9bba7d48d6a2","Type":"ContainerDied","Data":"578f35d8e0271bbd75028adc7c24d086b0f90dc84e95b7ca1d62dfea16e53168"} Dec 03 07:30:04 crc kubenswrapper[4818]: I1203 07:30:04.327281 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="578f35d8e0271bbd75028adc7c24d086b0f90dc84e95b7ca1d62dfea16e53168" Dec 03 07:30:04 crc kubenswrapper[4818]: I1203 07:30:04.326950 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412450-r76tm" Dec 03 07:30:04 crc kubenswrapper[4818]: I1203 07:30:04.731947 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv"] Dec 03 07:30:04 crc kubenswrapper[4818]: I1203 07:30:04.751071 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412405-4cctv"] Dec 03 07:30:06 crc kubenswrapper[4818]: I1203 07:30:06.748860 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7105fef-2722-4e9d-87ec-2553fbda781c" path="/var/lib/kubelet/pods/d7105fef-2722-4e9d-87ec-2553fbda781c/volumes" Dec 03 07:30:11 crc kubenswrapper[4818]: I1203 07:30:11.738079 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:30:11 crc kubenswrapper[4818]: E1203 07:30:11.738896 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:30:15 crc kubenswrapper[4818]: I1203 07:30:15.826635 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-xl5nj_47f0d000-e043-4e9e-b300-988f03341b42/cert-manager-controller/0.log" Dec 03 07:30:15 crc kubenswrapper[4818]: I1203 07:30:15.933201 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-cpcd5_c9bd0ac7-83ba-4f70-9458-0d8b88618456/cert-manager-cainjector/0.log" Dec 03 07:30:15 crc kubenswrapper[4818]: I1203 07:30:15.966307 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-tcn75_8f6981d3-9529-498d-935e-f99f533faa9e/cert-manager-webhook/0.log" Dec 03 07:30:23 crc kubenswrapper[4818]: I1203 07:30:23.738024 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:30:23 crc kubenswrapper[4818]: E1203 07:30:23.738896 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:30:27 crc kubenswrapper[4818]: I1203 07:30:27.828100 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-ncmjw_3c885e5f-3396-4927-90ef-878b255a6317/nmstate-console-plugin/0.log" Dec 03 07:30:28 crc kubenswrapper[4818]: I1203 07:30:28.015201 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-7hwmb_1d3707ea-c681-45e2-959f-336f11ad216d/nmstate-handler/0.log" Dec 03 07:30:28 crc kubenswrapper[4818]: I1203 07:30:28.088710 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-mrrpc_2ec105b0-3a09-4c8d-a494-95116b4329f2/kube-rbac-proxy/0.log" Dec 03 07:30:28 crc kubenswrapper[4818]: I1203 07:30:28.142241 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-mrrpc_2ec105b0-3a09-4c8d-a494-95116b4329f2/nmstate-metrics/0.log" Dec 03 07:30:28 crc kubenswrapper[4818]: I1203 07:30:28.327277 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-phlrm_160775bb-417d-4852-bc1f-77d6f8542fad/nmstate-operator/0.log" Dec 03 07:30:28 crc kubenswrapper[4818]: I1203 07:30:28.360537 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-lp854_fbefe4c2-3aea-4974-941e-74fc9448cf32/nmstate-webhook/0.log" Dec 03 07:30:38 crc kubenswrapper[4818]: I1203 07:30:38.737697 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:30:38 crc kubenswrapper[4818]: E1203 07:30:38.738469 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:30:41 crc kubenswrapper[4818]: I1203 07:30:41.743658 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-7nr97_bf628f10-0970-4651-8499-5302e78c8d6d/kube-rbac-proxy/0.log" Dec 03 07:30:41 crc kubenswrapper[4818]: I1203 07:30:41.915778 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-frr-files/0.log" Dec 03 07:30:41 crc kubenswrapper[4818]: I1203 07:30:41.935295 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-7nr97_bf628f10-0970-4651-8499-5302e78c8d6d/controller/0.log" Dec 03 07:30:42 crc kubenswrapper[4818]: I1203 07:30:42.139202 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-frr-files/0.log" Dec 03 07:30:42 crc kubenswrapper[4818]: I1203 07:30:42.179715 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-reloader/0.log" Dec 03 07:30:42 crc kubenswrapper[4818]: I1203 07:30:42.186927 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-metrics/0.log" Dec 03 07:30:42 crc kubenswrapper[4818]: I1203 07:30:42.216350 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-reloader/0.log" Dec 03 07:30:42 crc kubenswrapper[4818]: I1203 07:30:42.398235 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-frr-files/0.log" Dec 03 07:30:42 crc kubenswrapper[4818]: I1203 07:30:42.420872 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-metrics/0.log" Dec 03 07:30:42 crc kubenswrapper[4818]: I1203 07:30:42.439687 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-metrics/0.log" Dec 03 07:30:42 crc kubenswrapper[4818]: I1203 07:30:42.441087 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-reloader/0.log" Dec 03 07:30:42 crc kubenswrapper[4818]: I1203 07:30:42.579636 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-frr-files/0.log" Dec 03 07:30:42 crc kubenswrapper[4818]: I1203 07:30:42.599059 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-metrics/0.log" Dec 03 07:30:42 crc kubenswrapper[4818]: I1203 07:30:42.642185 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-reloader/0.log" Dec 03 07:30:42 crc kubenswrapper[4818]: I1203 07:30:42.655218 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/controller/0.log" Dec 03 07:30:42 crc kubenswrapper[4818]: I1203 07:30:42.802039 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/frr-metrics/0.log" Dec 03 07:30:42 crc kubenswrapper[4818]: I1203 07:30:42.836332 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/kube-rbac-proxy-frr/0.log" Dec 03 07:30:42 crc kubenswrapper[4818]: I1203 07:30:42.856150 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/kube-rbac-proxy/0.log" Dec 03 07:30:43 crc kubenswrapper[4818]: I1203 07:30:43.023639 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/reloader/0.log" Dec 03 07:30:43 crc kubenswrapper[4818]: I1203 07:30:43.089324 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-tkg7w_f55ffc30-37c1-45e8-8e63-c447ca4d6e96/frr-k8s-webhook-server/0.log" Dec 03 07:30:43 crc kubenswrapper[4818]: I1203 07:30:43.257336 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5545bb6cbd-gjkwk_d227f431-7cf6-43ed-85d0-6e44812df281/manager/0.log" Dec 03 07:30:43 crc kubenswrapper[4818]: I1203 07:30:43.450632 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-4m9n7_860c17a3-a17f-4ac3-9310-c633df611f6d/kube-rbac-proxy/0.log" Dec 03 07:30:43 crc kubenswrapper[4818]: I1203 07:30:43.495062 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-56f4c8658f-s8ghc_4563221c-8432-4038-b0e6-fdbac56b99e0/webhook-server/0.log" Dec 03 07:30:44 crc kubenswrapper[4818]: I1203 07:30:44.188213 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-4m9n7_860c17a3-a17f-4ac3-9310-c633df611f6d/speaker/0.log" Dec 03 07:30:44 crc kubenswrapper[4818]: I1203 07:30:44.254493 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/frr/0.log" Dec 03 07:30:52 crc kubenswrapper[4818]: I1203 07:30:52.749671 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:30:52 crc kubenswrapper[4818]: E1203 07:30:52.750950 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:30:55 crc kubenswrapper[4818]: I1203 07:30:55.718734 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5_37089b98-d9bb-4265-9e86-eeb0e4e0be99/util/0.log" Dec 03 07:30:55 crc kubenswrapper[4818]: I1203 07:30:55.920926 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5_37089b98-d9bb-4265-9e86-eeb0e4e0be99/util/0.log" Dec 03 07:30:55 crc kubenswrapper[4818]: I1203 07:30:55.921707 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5_37089b98-d9bb-4265-9e86-eeb0e4e0be99/pull/0.log" Dec 03 07:30:55 crc kubenswrapper[4818]: I1203 07:30:55.966411 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5_37089b98-d9bb-4265-9e86-eeb0e4e0be99/pull/0.log" Dec 03 07:30:56 crc kubenswrapper[4818]: I1203 07:30:56.221911 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5_37089b98-d9bb-4265-9e86-eeb0e4e0be99/pull/0.log" Dec 03 07:30:56 crc kubenswrapper[4818]: I1203 07:30:56.224269 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5_37089b98-d9bb-4265-9e86-eeb0e4e0be99/extract/0.log" Dec 03 07:30:56 crc kubenswrapper[4818]: I1203 07:30:56.224542 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5_37089b98-d9bb-4265-9e86-eeb0e4e0be99/util/0.log" Dec 03 07:30:56 crc kubenswrapper[4818]: I1203 07:30:56.367977 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm_9897d8d1-c815-45ee-9c90-0b1e34a71340/util/0.log" Dec 03 07:30:56 crc kubenswrapper[4818]: I1203 07:30:56.539388 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm_9897d8d1-c815-45ee-9c90-0b1e34a71340/util/0.log" Dec 03 07:30:56 crc kubenswrapper[4818]: I1203 07:30:56.560249 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm_9897d8d1-c815-45ee-9c90-0b1e34a71340/pull/0.log" Dec 03 07:30:56 crc kubenswrapper[4818]: I1203 07:30:56.573144 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm_9897d8d1-c815-45ee-9c90-0b1e34a71340/pull/0.log" Dec 03 07:30:56 crc kubenswrapper[4818]: I1203 07:30:56.729417 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm_9897d8d1-c815-45ee-9c90-0b1e34a71340/extract/0.log" Dec 03 07:30:56 crc kubenswrapper[4818]: I1203 07:30:56.743228 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm_9897d8d1-c815-45ee-9c90-0b1e34a71340/util/0.log" Dec 03 07:30:56 crc kubenswrapper[4818]: I1203 07:30:56.744151 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm_9897d8d1-c815-45ee-9c90-0b1e34a71340/pull/0.log" Dec 03 07:30:56 crc kubenswrapper[4818]: I1203 07:30:56.906253 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2kmpd_93370008-4229-4f0f-be59-8385406d7cd1/extract-utilities/0.log" Dec 03 07:30:57 crc kubenswrapper[4818]: I1203 07:30:57.098853 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2kmpd_93370008-4229-4f0f-be59-8385406d7cd1/extract-content/0.log" Dec 03 07:30:57 crc kubenswrapper[4818]: I1203 07:30:57.099249 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2kmpd_93370008-4229-4f0f-be59-8385406d7cd1/extract-utilities/0.log" Dec 03 07:30:57 crc kubenswrapper[4818]: I1203 07:30:57.127049 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2kmpd_93370008-4229-4f0f-be59-8385406d7cd1/extract-content/0.log" Dec 03 07:30:57 crc kubenswrapper[4818]: I1203 07:30:57.302634 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2kmpd_93370008-4229-4f0f-be59-8385406d7cd1/extract-content/0.log" Dec 03 07:30:57 crc kubenswrapper[4818]: I1203 07:30:57.314917 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2kmpd_93370008-4229-4f0f-be59-8385406d7cd1/extract-utilities/0.log" Dec 03 07:30:57 crc kubenswrapper[4818]: I1203 07:30:57.500248 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mwgrm_57e20fce-c349-473f-8866-b194da09c27d/extract-utilities/0.log" Dec 03 07:30:57 crc kubenswrapper[4818]: I1203 07:30:57.703020 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mwgrm_57e20fce-c349-473f-8866-b194da09c27d/extract-content/0.log" Dec 03 07:30:57 crc kubenswrapper[4818]: I1203 07:30:57.705670 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mwgrm_57e20fce-c349-473f-8866-b194da09c27d/extract-utilities/0.log" Dec 03 07:30:57 crc kubenswrapper[4818]: I1203 07:30:57.713697 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mwgrm_57e20fce-c349-473f-8866-b194da09c27d/extract-content/0.log" Dec 03 07:30:57 crc kubenswrapper[4818]: I1203 07:30:57.887313 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mwgrm_57e20fce-c349-473f-8866-b194da09c27d/extract-content/0.log" Dec 03 07:30:57 crc kubenswrapper[4818]: I1203 07:30:57.934439 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mwgrm_57e20fce-c349-473f-8866-b194da09c27d/extract-utilities/0.log" Dec 03 07:30:58 crc kubenswrapper[4818]: I1203 07:30:58.140367 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hl66m_a89a1dfa-1797-48f8-bd2d-f968c523e8df/marketplace-operator/0.log" Dec 03 07:30:58 crc kubenswrapper[4818]: I1203 07:30:58.407521 4818 scope.go:117] "RemoveContainer" containerID="54145502f815f6579c84f12f671d0ddd2d4ce8b94b904d29f5178b57ea8a4da3" Dec 03 07:30:58 crc kubenswrapper[4818]: I1203 07:30:58.466613 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dm6xl_b2a86a74-f04c-49ed-a900-33a852b794a3/extract-utilities/0.log" Dec 03 07:30:58 crc kubenswrapper[4818]: I1203 07:30:58.640012 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dm6xl_b2a86a74-f04c-49ed-a900-33a852b794a3/extract-utilities/0.log" Dec 03 07:30:58 crc kubenswrapper[4818]: I1203 07:30:58.653252 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2kmpd_93370008-4229-4f0f-be59-8385406d7cd1/registry-server/0.log" Dec 03 07:30:58 crc kubenswrapper[4818]: I1203 07:30:58.666860 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dm6xl_b2a86a74-f04c-49ed-a900-33a852b794a3/extract-content/0.log" Dec 03 07:30:58 crc kubenswrapper[4818]: I1203 07:30:58.761542 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mwgrm_57e20fce-c349-473f-8866-b194da09c27d/registry-server/0.log" Dec 03 07:30:58 crc kubenswrapper[4818]: I1203 07:30:58.822491 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dm6xl_b2a86a74-f04c-49ed-a900-33a852b794a3/extract-content/0.log" Dec 03 07:30:59 crc kubenswrapper[4818]: I1203 07:30:59.033675 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dm6xl_b2a86a74-f04c-49ed-a900-33a852b794a3/extract-content/0.log" Dec 03 07:30:59 crc kubenswrapper[4818]: I1203 07:30:59.109797 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dm6xl_b2a86a74-f04c-49ed-a900-33a852b794a3/extract-utilities/0.log" Dec 03 07:30:59 crc kubenswrapper[4818]: I1203 07:30:59.235169 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dm6xl_b2a86a74-f04c-49ed-a900-33a852b794a3/registry-server/0.log" Dec 03 07:30:59 crc kubenswrapper[4818]: I1203 07:30:59.260918 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hchtl_761a2fb4-d124-4db1-9924-0cd3d9546fa5/extract-utilities/0.log" Dec 03 07:30:59 crc kubenswrapper[4818]: I1203 07:30:59.426195 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hchtl_761a2fb4-d124-4db1-9924-0cd3d9546fa5/extract-content/0.log" Dec 03 07:30:59 crc kubenswrapper[4818]: I1203 07:30:59.431024 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hchtl_761a2fb4-d124-4db1-9924-0cd3d9546fa5/extract-utilities/0.log" Dec 03 07:30:59 crc kubenswrapper[4818]: I1203 07:30:59.456163 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hchtl_761a2fb4-d124-4db1-9924-0cd3d9546fa5/extract-content/0.log" Dec 03 07:30:59 crc kubenswrapper[4818]: I1203 07:30:59.610924 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hchtl_761a2fb4-d124-4db1-9924-0cd3d9546fa5/extract-utilities/0.log" Dec 03 07:30:59 crc kubenswrapper[4818]: I1203 07:30:59.626964 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hchtl_761a2fb4-d124-4db1-9924-0cd3d9546fa5/extract-content/0.log" Dec 03 07:31:00 crc kubenswrapper[4818]: I1203 07:31:00.127754 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hchtl_761a2fb4-d124-4db1-9924-0cd3d9546fa5/registry-server/0.log" Dec 03 07:31:03 crc kubenswrapper[4818]: I1203 07:31:03.738363 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:31:03 crc kubenswrapper[4818]: E1203 07:31:03.740343 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:31:18 crc kubenswrapper[4818]: I1203 07:31:18.739145 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:31:18 crc kubenswrapper[4818]: E1203 07:31:18.740211 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:31:30 crc kubenswrapper[4818]: I1203 07:31:30.737627 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:31:30 crc kubenswrapper[4818]: E1203 07:31:30.738413 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:31:42 crc kubenswrapper[4818]: I1203 07:31:42.748657 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:31:42 crc kubenswrapper[4818]: E1203 07:31:42.749739 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:31:57 crc kubenswrapper[4818]: I1203 07:31:57.737576 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:31:57 crc kubenswrapper[4818]: E1203 07:31:57.738427 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:32:08 crc kubenswrapper[4818]: I1203 07:32:08.738008 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:32:08 crc kubenswrapper[4818]: E1203 07:32:08.738737 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:32:22 crc kubenswrapper[4818]: I1203 07:32:22.745431 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:32:22 crc kubenswrapper[4818]: E1203 07:32:22.746521 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:32:31 crc kubenswrapper[4818]: I1203 07:32:31.216332 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-7954d7ccff-th9lw" podUID="0ba50d38-959d-49f4-a98a-6280a78661e4" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 03 07:32:37 crc kubenswrapper[4818]: I1203 07:32:37.739052 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:32:37 crc kubenswrapper[4818]: E1203 07:32:37.740098 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:32:40 crc kubenswrapper[4818]: I1203 07:32:40.812814 4818 generic.go:334] "Generic (PLEG): container finished" podID="18e6fb67-550a-4057-a619-6ccfb3707649" containerID="dcefd7bbb59c7b1a55fb0d1e9f927d7a0357766e6cc5bdc543da12480e666e82" exitCode=0 Dec 03 07:32:40 crc kubenswrapper[4818]: I1203 07:32:40.812854 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ffrv8/must-gather-2cb84" event={"ID":"18e6fb67-550a-4057-a619-6ccfb3707649","Type":"ContainerDied","Data":"dcefd7bbb59c7b1a55fb0d1e9f927d7a0357766e6cc5bdc543da12480e666e82"} Dec 03 07:32:40 crc kubenswrapper[4818]: I1203 07:32:40.813861 4818 scope.go:117] "RemoveContainer" containerID="dcefd7bbb59c7b1a55fb0d1e9f927d7a0357766e6cc5bdc543da12480e666e82" Dec 03 07:32:41 crc kubenswrapper[4818]: I1203 07:32:41.209482 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ffrv8_must-gather-2cb84_18e6fb67-550a-4057-a619-6ccfb3707649/gather/0.log" Dec 03 07:32:50 crc kubenswrapper[4818]: I1203 07:32:50.739160 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:32:51 crc kubenswrapper[4818]: I1203 07:32:51.548364 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ffrv8/must-gather-2cb84"] Dec 03 07:32:51 crc kubenswrapper[4818]: I1203 07:32:51.548946 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-ffrv8/must-gather-2cb84" podUID="18e6fb67-550a-4057-a619-6ccfb3707649" containerName="copy" containerID="cri-o://68c62b78adde405f1777d96a033669e55a0ef61cb60ab62dbefaf335e3513038" gracePeriod=2 Dec 03 07:32:51 crc kubenswrapper[4818]: I1203 07:32:51.556639 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ffrv8/must-gather-2cb84"] Dec 03 07:32:51 crc kubenswrapper[4818]: I1203 07:32:51.904307 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"3696ecd804b3cc54d22f8baf9c3f030c9a0933db13e3f5ad4cac7c6ea9cfefef"} Dec 03 07:32:51 crc kubenswrapper[4818]: I1203 07:32:51.907256 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ffrv8_must-gather-2cb84_18e6fb67-550a-4057-a619-6ccfb3707649/copy/0.log" Dec 03 07:32:51 crc kubenswrapper[4818]: I1203 07:32:51.907702 4818 generic.go:334] "Generic (PLEG): container finished" podID="18e6fb67-550a-4057-a619-6ccfb3707649" containerID="68c62b78adde405f1777d96a033669e55a0ef61cb60ab62dbefaf335e3513038" exitCode=143 Dec 03 07:32:52 crc kubenswrapper[4818]: I1203 07:32:52.528642 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ffrv8_must-gather-2cb84_18e6fb67-550a-4057-a619-6ccfb3707649/copy/0.log" Dec 03 07:32:52 crc kubenswrapper[4818]: I1203 07:32:52.529494 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffrv8/must-gather-2cb84" Dec 03 07:32:52 crc kubenswrapper[4818]: I1203 07:32:52.672943 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4v5m\" (UniqueName: \"kubernetes.io/projected/18e6fb67-550a-4057-a619-6ccfb3707649-kube-api-access-l4v5m\") pod \"18e6fb67-550a-4057-a619-6ccfb3707649\" (UID: \"18e6fb67-550a-4057-a619-6ccfb3707649\") " Dec 03 07:32:52 crc kubenswrapper[4818]: I1203 07:32:52.673065 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/18e6fb67-550a-4057-a619-6ccfb3707649-must-gather-output\") pod \"18e6fb67-550a-4057-a619-6ccfb3707649\" (UID: \"18e6fb67-550a-4057-a619-6ccfb3707649\") " Dec 03 07:32:52 crc kubenswrapper[4818]: I1203 07:32:52.679284 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18e6fb67-550a-4057-a619-6ccfb3707649-kube-api-access-l4v5m" (OuterVolumeSpecName: "kube-api-access-l4v5m") pod "18e6fb67-550a-4057-a619-6ccfb3707649" (UID: "18e6fb67-550a-4057-a619-6ccfb3707649"). InnerVolumeSpecName "kube-api-access-l4v5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:32:52 crc kubenswrapper[4818]: I1203 07:32:52.775103 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4v5m\" (UniqueName: \"kubernetes.io/projected/18e6fb67-550a-4057-a619-6ccfb3707649-kube-api-access-l4v5m\") on node \"crc\" DevicePath \"\"" Dec 03 07:32:52 crc kubenswrapper[4818]: I1203 07:32:52.818291 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18e6fb67-550a-4057-a619-6ccfb3707649-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "18e6fb67-550a-4057-a619-6ccfb3707649" (UID: "18e6fb67-550a-4057-a619-6ccfb3707649"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:32:52 crc kubenswrapper[4818]: I1203 07:32:52.877120 4818 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/18e6fb67-550a-4057-a619-6ccfb3707649-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 07:32:52 crc kubenswrapper[4818]: I1203 07:32:52.917848 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ffrv8_must-gather-2cb84_18e6fb67-550a-4057-a619-6ccfb3707649/copy/0.log" Dec 03 07:32:52 crc kubenswrapper[4818]: I1203 07:32:52.918613 4818 scope.go:117] "RemoveContainer" containerID="68c62b78adde405f1777d96a033669e55a0ef61cb60ab62dbefaf335e3513038" Dec 03 07:32:52 crc kubenswrapper[4818]: I1203 07:32:52.918656 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ffrv8/must-gather-2cb84" Dec 03 07:32:52 crc kubenswrapper[4818]: I1203 07:32:52.941520 4818 scope.go:117] "RemoveContainer" containerID="dcefd7bbb59c7b1a55fb0d1e9f927d7a0357766e6cc5bdc543da12480e666e82" Dec 03 07:32:54 crc kubenswrapper[4818]: I1203 07:32:54.752531 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18e6fb67-550a-4057-a619-6ccfb3707649" path="/var/lib/kubelet/pods/18e6fb67-550a-4057-a619-6ccfb3707649/volumes" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.243495 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l66pl"] Dec 03 07:35:10 crc kubenswrapper[4818]: E1203 07:35:10.244858 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18e6fb67-550a-4057-a619-6ccfb3707649" containerName="copy" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.244883 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="18e6fb67-550a-4057-a619-6ccfb3707649" containerName="copy" Dec 03 07:35:10 crc kubenswrapper[4818]: E1203 07:35:10.244927 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9c72d04-9a02-4a2d-b458-9bba7d48d6a2" containerName="collect-profiles" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.244941 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9c72d04-9a02-4a2d-b458-9bba7d48d6a2" containerName="collect-profiles" Dec 03 07:35:10 crc kubenswrapper[4818]: E1203 07:35:10.245003 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18e6fb67-550a-4057-a619-6ccfb3707649" containerName="gather" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.245017 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="18e6fb67-550a-4057-a619-6ccfb3707649" containerName="gather" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.245434 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="18e6fb67-550a-4057-a619-6ccfb3707649" containerName="copy" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.245473 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="18e6fb67-550a-4057-a619-6ccfb3707649" containerName="gather" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.245512 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9c72d04-9a02-4a2d-b458-9bba7d48d6a2" containerName="collect-profiles" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.248333 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l66pl" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.262540 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l66pl"] Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.313120 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/281a83c0-1e5e-4dac-b022-9904d0ac439e-utilities\") pod \"certified-operators-l66pl\" (UID: \"281a83c0-1e5e-4dac-b022-9904d0ac439e\") " pod="openshift-marketplace/certified-operators-l66pl" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.313191 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9kv6\" (UniqueName: \"kubernetes.io/projected/281a83c0-1e5e-4dac-b022-9904d0ac439e-kube-api-access-z9kv6\") pod \"certified-operators-l66pl\" (UID: \"281a83c0-1e5e-4dac-b022-9904d0ac439e\") " pod="openshift-marketplace/certified-operators-l66pl" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.313250 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/281a83c0-1e5e-4dac-b022-9904d0ac439e-catalog-content\") pod \"certified-operators-l66pl\" (UID: \"281a83c0-1e5e-4dac-b022-9904d0ac439e\") " pod="openshift-marketplace/certified-operators-l66pl" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.414539 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/281a83c0-1e5e-4dac-b022-9904d0ac439e-utilities\") pod \"certified-operators-l66pl\" (UID: \"281a83c0-1e5e-4dac-b022-9904d0ac439e\") " pod="openshift-marketplace/certified-operators-l66pl" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.414590 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9kv6\" (UniqueName: \"kubernetes.io/projected/281a83c0-1e5e-4dac-b022-9904d0ac439e-kube-api-access-z9kv6\") pod \"certified-operators-l66pl\" (UID: \"281a83c0-1e5e-4dac-b022-9904d0ac439e\") " pod="openshift-marketplace/certified-operators-l66pl" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.414627 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/281a83c0-1e5e-4dac-b022-9904d0ac439e-catalog-content\") pod \"certified-operators-l66pl\" (UID: \"281a83c0-1e5e-4dac-b022-9904d0ac439e\") " pod="openshift-marketplace/certified-operators-l66pl" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.415178 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/281a83c0-1e5e-4dac-b022-9904d0ac439e-utilities\") pod \"certified-operators-l66pl\" (UID: \"281a83c0-1e5e-4dac-b022-9904d0ac439e\") " pod="openshift-marketplace/certified-operators-l66pl" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.415215 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/281a83c0-1e5e-4dac-b022-9904d0ac439e-catalog-content\") pod \"certified-operators-l66pl\" (UID: \"281a83c0-1e5e-4dac-b022-9904d0ac439e\") " pod="openshift-marketplace/certified-operators-l66pl" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.437842 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9kv6\" (UniqueName: \"kubernetes.io/projected/281a83c0-1e5e-4dac-b022-9904d0ac439e-kube-api-access-z9kv6\") pod \"certified-operators-l66pl\" (UID: \"281a83c0-1e5e-4dac-b022-9904d0ac439e\") " pod="openshift-marketplace/certified-operators-l66pl" Dec 03 07:35:10 crc kubenswrapper[4818]: I1203 07:35:10.580050 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l66pl" Dec 03 07:35:11 crc kubenswrapper[4818]: I1203 07:35:11.129219 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l66pl"] Dec 03 07:35:11 crc kubenswrapper[4818]: I1203 07:35:11.220058 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l66pl" event={"ID":"281a83c0-1e5e-4dac-b022-9904d0ac439e","Type":"ContainerStarted","Data":"9824faadd4c4d81b0a00b7aa33a31a623fc26de794d373c2e35468b219467eb6"} Dec 03 07:35:12 crc kubenswrapper[4818]: I1203 07:35:12.232489 4818 generic.go:334] "Generic (PLEG): container finished" podID="281a83c0-1e5e-4dac-b022-9904d0ac439e" containerID="7882733c1d86342245ff2ce06a9b44a8acb4191acf6847ab421556af15c44f25" exitCode=0 Dec 03 07:35:12 crc kubenswrapper[4818]: I1203 07:35:12.232569 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l66pl" event={"ID":"281a83c0-1e5e-4dac-b022-9904d0ac439e","Type":"ContainerDied","Data":"7882733c1d86342245ff2ce06a9b44a8acb4191acf6847ab421556af15c44f25"} Dec 03 07:35:12 crc kubenswrapper[4818]: I1203 07:35:12.235411 4818 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 07:35:12 crc kubenswrapper[4818]: I1203 07:35:12.433113 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xcqtm"] Dec 03 07:35:12 crc kubenswrapper[4818]: I1203 07:35:12.435249 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xcqtm" Dec 03 07:35:12 crc kubenswrapper[4818]: I1203 07:35:12.441855 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xcqtm"] Dec 03 07:35:12 crc kubenswrapper[4818]: I1203 07:35:12.572197 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f077309-37ed-4977-8638-86dacef46390-catalog-content\") pod \"redhat-marketplace-xcqtm\" (UID: \"0f077309-37ed-4977-8638-86dacef46390\") " pod="openshift-marketplace/redhat-marketplace-xcqtm" Dec 03 07:35:12 crc kubenswrapper[4818]: I1203 07:35:12.572242 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrq6c\" (UniqueName: \"kubernetes.io/projected/0f077309-37ed-4977-8638-86dacef46390-kube-api-access-nrq6c\") pod \"redhat-marketplace-xcqtm\" (UID: \"0f077309-37ed-4977-8638-86dacef46390\") " pod="openshift-marketplace/redhat-marketplace-xcqtm" Dec 03 07:35:12 crc kubenswrapper[4818]: I1203 07:35:12.572314 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f077309-37ed-4977-8638-86dacef46390-utilities\") pod \"redhat-marketplace-xcqtm\" (UID: \"0f077309-37ed-4977-8638-86dacef46390\") " pod="openshift-marketplace/redhat-marketplace-xcqtm" Dec 03 07:35:12 crc kubenswrapper[4818]: I1203 07:35:12.673777 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f077309-37ed-4977-8638-86dacef46390-utilities\") pod \"redhat-marketplace-xcqtm\" (UID: \"0f077309-37ed-4977-8638-86dacef46390\") " pod="openshift-marketplace/redhat-marketplace-xcqtm" Dec 03 07:35:12 crc kubenswrapper[4818]: I1203 07:35:12.673995 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f077309-37ed-4977-8638-86dacef46390-catalog-content\") pod \"redhat-marketplace-xcqtm\" (UID: \"0f077309-37ed-4977-8638-86dacef46390\") " pod="openshift-marketplace/redhat-marketplace-xcqtm" Dec 03 07:35:12 crc kubenswrapper[4818]: I1203 07:35:12.674027 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrq6c\" (UniqueName: \"kubernetes.io/projected/0f077309-37ed-4977-8638-86dacef46390-kube-api-access-nrq6c\") pod \"redhat-marketplace-xcqtm\" (UID: \"0f077309-37ed-4977-8638-86dacef46390\") " pod="openshift-marketplace/redhat-marketplace-xcqtm" Dec 03 07:35:12 crc kubenswrapper[4818]: I1203 07:35:12.674455 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f077309-37ed-4977-8638-86dacef46390-utilities\") pod \"redhat-marketplace-xcqtm\" (UID: \"0f077309-37ed-4977-8638-86dacef46390\") " pod="openshift-marketplace/redhat-marketplace-xcqtm" Dec 03 07:35:12 crc kubenswrapper[4818]: I1203 07:35:12.674569 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f077309-37ed-4977-8638-86dacef46390-catalog-content\") pod \"redhat-marketplace-xcqtm\" (UID: \"0f077309-37ed-4977-8638-86dacef46390\") " pod="openshift-marketplace/redhat-marketplace-xcqtm" Dec 03 07:35:12 crc kubenswrapper[4818]: I1203 07:35:12.695646 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrq6c\" (UniqueName: \"kubernetes.io/projected/0f077309-37ed-4977-8638-86dacef46390-kube-api-access-nrq6c\") pod \"redhat-marketplace-xcqtm\" (UID: \"0f077309-37ed-4977-8638-86dacef46390\") " pod="openshift-marketplace/redhat-marketplace-xcqtm" Dec 03 07:35:12 crc kubenswrapper[4818]: I1203 07:35:12.765924 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xcqtm" Dec 03 07:35:13 crc kubenswrapper[4818]: W1203 07:35:13.258146 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f077309_37ed_4977_8638_86dacef46390.slice/crio-affb599f4cd96a888cca21dc0c45e8637a9bf3aed92b8937d0e47f69b9955c73 WatchSource:0}: Error finding container affb599f4cd96a888cca21dc0c45e8637a9bf3aed92b8937d0e47f69b9955c73: Status 404 returned error can't find the container with id affb599f4cd96a888cca21dc0c45e8637a9bf3aed92b8937d0e47f69b9955c73 Dec 03 07:35:13 crc kubenswrapper[4818]: I1203 07:35:13.258342 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l66pl" event={"ID":"281a83c0-1e5e-4dac-b022-9904d0ac439e","Type":"ContainerStarted","Data":"585d8d2cf9fdcf91e523dd19584c7c2dbba568e97fe2a5258256c2fc8af9901b"} Dec 03 07:35:13 crc kubenswrapper[4818]: I1203 07:35:13.280975 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xcqtm"] Dec 03 07:35:13 crc kubenswrapper[4818]: I1203 07:35:13.302086 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:35:13 crc kubenswrapper[4818]: I1203 07:35:13.302220 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:35:13 crc kubenswrapper[4818]: I1203 07:35:13.436376 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8slz4"] Dec 03 07:35:13 crc kubenswrapper[4818]: I1203 07:35:13.442006 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8slz4" Dec 03 07:35:13 crc kubenswrapper[4818]: I1203 07:35:13.451506 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8slz4"] Dec 03 07:35:13 crc kubenswrapper[4818]: I1203 07:35:13.490458 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-catalog-content\") pod \"redhat-operators-8slz4\" (UID: \"ef736fcb-502c-4a07-8b2f-cc0bc77297e9\") " pod="openshift-marketplace/redhat-operators-8slz4" Dec 03 07:35:13 crc kubenswrapper[4818]: I1203 07:35:13.490575 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd54x\" (UniqueName: \"kubernetes.io/projected/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-kube-api-access-wd54x\") pod \"redhat-operators-8slz4\" (UID: \"ef736fcb-502c-4a07-8b2f-cc0bc77297e9\") " pod="openshift-marketplace/redhat-operators-8slz4" Dec 03 07:35:13 crc kubenswrapper[4818]: I1203 07:35:13.490637 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-utilities\") pod \"redhat-operators-8slz4\" (UID: \"ef736fcb-502c-4a07-8b2f-cc0bc77297e9\") " pod="openshift-marketplace/redhat-operators-8slz4" Dec 03 07:35:13 crc kubenswrapper[4818]: I1203 07:35:13.592157 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd54x\" (UniqueName: \"kubernetes.io/projected/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-kube-api-access-wd54x\") pod \"redhat-operators-8slz4\" (UID: \"ef736fcb-502c-4a07-8b2f-cc0bc77297e9\") " pod="openshift-marketplace/redhat-operators-8slz4" Dec 03 07:35:13 crc kubenswrapper[4818]: I1203 07:35:13.592257 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-utilities\") pod \"redhat-operators-8slz4\" (UID: \"ef736fcb-502c-4a07-8b2f-cc0bc77297e9\") " pod="openshift-marketplace/redhat-operators-8slz4" Dec 03 07:35:13 crc kubenswrapper[4818]: I1203 07:35:13.592341 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-catalog-content\") pod \"redhat-operators-8slz4\" (UID: \"ef736fcb-502c-4a07-8b2f-cc0bc77297e9\") " pod="openshift-marketplace/redhat-operators-8slz4" Dec 03 07:35:13 crc kubenswrapper[4818]: I1203 07:35:13.592861 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-utilities\") pod \"redhat-operators-8slz4\" (UID: \"ef736fcb-502c-4a07-8b2f-cc0bc77297e9\") " pod="openshift-marketplace/redhat-operators-8slz4" Dec 03 07:35:13 crc kubenswrapper[4818]: I1203 07:35:13.592882 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-catalog-content\") pod \"redhat-operators-8slz4\" (UID: \"ef736fcb-502c-4a07-8b2f-cc0bc77297e9\") " pod="openshift-marketplace/redhat-operators-8slz4" Dec 03 07:35:13 crc kubenswrapper[4818]: I1203 07:35:13.614326 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd54x\" (UniqueName: \"kubernetes.io/projected/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-kube-api-access-wd54x\") pod \"redhat-operators-8slz4\" (UID: \"ef736fcb-502c-4a07-8b2f-cc0bc77297e9\") " pod="openshift-marketplace/redhat-operators-8slz4" Dec 03 07:35:13 crc kubenswrapper[4818]: I1203 07:35:13.798560 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8slz4" Dec 03 07:35:14 crc kubenswrapper[4818]: I1203 07:35:14.237411 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8slz4"] Dec 03 07:35:14 crc kubenswrapper[4818]: I1203 07:35:14.272540 4818 generic.go:334] "Generic (PLEG): container finished" podID="0f077309-37ed-4977-8638-86dacef46390" containerID="be879c2f962530bd21ac89b79c90b00f60b27d12ee40d5914efedb644b3c2550" exitCode=0 Dec 03 07:35:14 crc kubenswrapper[4818]: I1203 07:35:14.272606 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xcqtm" event={"ID":"0f077309-37ed-4977-8638-86dacef46390","Type":"ContainerDied","Data":"be879c2f962530bd21ac89b79c90b00f60b27d12ee40d5914efedb644b3c2550"} Dec 03 07:35:14 crc kubenswrapper[4818]: I1203 07:35:14.272642 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xcqtm" event={"ID":"0f077309-37ed-4977-8638-86dacef46390","Type":"ContainerStarted","Data":"affb599f4cd96a888cca21dc0c45e8637a9bf3aed92b8937d0e47f69b9955c73"} Dec 03 07:35:14 crc kubenswrapper[4818]: I1203 07:35:14.275918 4818 generic.go:334] "Generic (PLEG): container finished" podID="281a83c0-1e5e-4dac-b022-9904d0ac439e" containerID="585d8d2cf9fdcf91e523dd19584c7c2dbba568e97fe2a5258256c2fc8af9901b" exitCode=0 Dec 03 07:35:14 crc kubenswrapper[4818]: I1203 07:35:14.275966 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l66pl" event={"ID":"281a83c0-1e5e-4dac-b022-9904d0ac439e","Type":"ContainerDied","Data":"585d8d2cf9fdcf91e523dd19584c7c2dbba568e97fe2a5258256c2fc8af9901b"} Dec 03 07:35:14 crc kubenswrapper[4818]: W1203 07:35:14.360900 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef736fcb_502c_4a07_8b2f_cc0bc77297e9.slice/crio-4caa594a06d9254ec48cafb41ed1560824e197b5bbd197793320f5d0cc0a33ed WatchSource:0}: Error finding container 4caa594a06d9254ec48cafb41ed1560824e197b5bbd197793320f5d0cc0a33ed: Status 404 returned error can't find the container with id 4caa594a06d9254ec48cafb41ed1560824e197b5bbd197793320f5d0cc0a33ed Dec 03 07:35:15 crc kubenswrapper[4818]: I1203 07:35:15.297544 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l66pl" event={"ID":"281a83c0-1e5e-4dac-b022-9904d0ac439e","Type":"ContainerStarted","Data":"f82ca608b54cfa94963278532a14affc4f8e2e407466a7aece9d87cc5fd9620d"} Dec 03 07:35:15 crc kubenswrapper[4818]: I1203 07:35:15.299472 4818 generic.go:334] "Generic (PLEG): container finished" podID="ef736fcb-502c-4a07-8b2f-cc0bc77297e9" containerID="fefc67c0bfc45bf4f8dea70aaaba838229f346a629d064afdd99fe4b98d58ca4" exitCode=0 Dec 03 07:35:15 crc kubenswrapper[4818]: I1203 07:35:15.299529 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8slz4" event={"ID":"ef736fcb-502c-4a07-8b2f-cc0bc77297e9","Type":"ContainerDied","Data":"fefc67c0bfc45bf4f8dea70aaaba838229f346a629d064afdd99fe4b98d58ca4"} Dec 03 07:35:15 crc kubenswrapper[4818]: I1203 07:35:15.299564 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8slz4" event={"ID":"ef736fcb-502c-4a07-8b2f-cc0bc77297e9","Type":"ContainerStarted","Data":"4caa594a06d9254ec48cafb41ed1560824e197b5bbd197793320f5d0cc0a33ed"} Dec 03 07:35:15 crc kubenswrapper[4818]: I1203 07:35:15.328147 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l66pl" podStartSLOduration=2.770043004 podStartE2EDuration="5.328120062s" podCreationTimestamp="2025-12-03 07:35:10 +0000 UTC" firstStartedPulling="2025-12-03 07:35:12.234845887 +0000 UTC m=+4069.926454669" lastFinishedPulling="2025-12-03 07:35:14.792922975 +0000 UTC m=+4072.484531727" observedRunningTime="2025-12-03 07:35:15.314061422 +0000 UTC m=+4073.005670184" watchObservedRunningTime="2025-12-03 07:35:15.328120062 +0000 UTC m=+4073.019728834" Dec 03 07:35:16 crc kubenswrapper[4818]: I1203 07:35:16.307942 4818 generic.go:334] "Generic (PLEG): container finished" podID="0f077309-37ed-4977-8638-86dacef46390" containerID="0489d5b2238e098100040a5c786647acd45cbc3a30242d4dcc58a4e5b3241104" exitCode=0 Dec 03 07:35:16 crc kubenswrapper[4818]: I1203 07:35:16.308297 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xcqtm" event={"ID":"0f077309-37ed-4977-8638-86dacef46390","Type":"ContainerDied","Data":"0489d5b2238e098100040a5c786647acd45cbc3a30242d4dcc58a4e5b3241104"} Dec 03 07:35:17 crc kubenswrapper[4818]: I1203 07:35:17.323537 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8slz4" event={"ID":"ef736fcb-502c-4a07-8b2f-cc0bc77297e9","Type":"ContainerStarted","Data":"cbe32f7471f49b87914db9ec53b4114168782924d378d14ccc2d211a4b1e8610"} Dec 03 07:35:17 crc kubenswrapper[4818]: I1203 07:35:17.327669 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xcqtm" event={"ID":"0f077309-37ed-4977-8638-86dacef46390","Type":"ContainerStarted","Data":"2e9f133b313e955e6848cac51386d402692adbae72697deec6d010220f6ec485"} Dec 03 07:35:17 crc kubenswrapper[4818]: I1203 07:35:17.373719 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xcqtm" podStartSLOduration=3.004208574 podStartE2EDuration="5.373698118s" podCreationTimestamp="2025-12-03 07:35:12 +0000 UTC" firstStartedPulling="2025-12-03 07:35:14.359137976 +0000 UTC m=+4072.050746738" lastFinishedPulling="2025-12-03 07:35:16.72862753 +0000 UTC m=+4074.420236282" observedRunningTime="2025-12-03 07:35:17.369449552 +0000 UTC m=+4075.061058324" watchObservedRunningTime="2025-12-03 07:35:17.373698118 +0000 UTC m=+4075.065306870" Dec 03 07:35:20 crc kubenswrapper[4818]: I1203 07:35:20.356231 4818 generic.go:334] "Generic (PLEG): container finished" podID="ef736fcb-502c-4a07-8b2f-cc0bc77297e9" containerID="cbe32f7471f49b87914db9ec53b4114168782924d378d14ccc2d211a4b1e8610" exitCode=0 Dec 03 07:35:20 crc kubenswrapper[4818]: I1203 07:35:20.356344 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8slz4" event={"ID":"ef736fcb-502c-4a07-8b2f-cc0bc77297e9","Type":"ContainerDied","Data":"cbe32f7471f49b87914db9ec53b4114168782924d378d14ccc2d211a4b1e8610"} Dec 03 07:35:20 crc kubenswrapper[4818]: I1203 07:35:20.580524 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l66pl" Dec 03 07:35:20 crc kubenswrapper[4818]: I1203 07:35:20.580594 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l66pl" Dec 03 07:35:20 crc kubenswrapper[4818]: I1203 07:35:20.662523 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l66pl" Dec 03 07:35:21 crc kubenswrapper[4818]: I1203 07:35:21.432515 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l66pl" Dec 03 07:35:22 crc kubenswrapper[4818]: I1203 07:35:22.217271 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l66pl"] Dec 03 07:35:22 crc kubenswrapper[4818]: I1203 07:35:22.767154 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xcqtm" Dec 03 07:35:22 crc kubenswrapper[4818]: I1203 07:35:22.767421 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xcqtm" Dec 03 07:35:22 crc kubenswrapper[4818]: I1203 07:35:22.825766 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xcqtm" Dec 03 07:35:23 crc kubenswrapper[4818]: I1203 07:35:23.460445 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8slz4" event={"ID":"ef736fcb-502c-4a07-8b2f-cc0bc77297e9","Type":"ContainerStarted","Data":"eaa2db839379be9567bac92037002c79dec255f9f47316c13a95a7c5f5dc92ad"} Dec 03 07:35:23 crc kubenswrapper[4818]: I1203 07:35:23.460759 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l66pl" podUID="281a83c0-1e5e-4dac-b022-9904d0ac439e" containerName="registry-server" containerID="cri-o://f82ca608b54cfa94963278532a14affc4f8e2e407466a7aece9d87cc5fd9620d" gracePeriod=2 Dec 03 07:35:23 crc kubenswrapper[4818]: I1203 07:35:23.488794 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8slz4" podStartSLOduration=2.895229076 podStartE2EDuration="10.488774134s" podCreationTimestamp="2025-12-03 07:35:13 +0000 UTC" firstStartedPulling="2025-12-03 07:35:15.301023477 +0000 UTC m=+4072.992632239" lastFinishedPulling="2025-12-03 07:35:22.894568545 +0000 UTC m=+4080.586177297" observedRunningTime="2025-12-03 07:35:23.480685832 +0000 UTC m=+4081.172294584" watchObservedRunningTime="2025-12-03 07:35:23.488774134 +0000 UTC m=+4081.180382886" Dec 03 07:35:23 crc kubenswrapper[4818]: I1203 07:35:23.514752 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xcqtm" Dec 03 07:35:23 crc kubenswrapper[4818]: I1203 07:35:23.799639 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8slz4" Dec 03 07:35:23 crc kubenswrapper[4818]: I1203 07:35:23.800008 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8slz4" Dec 03 07:35:23 crc kubenswrapper[4818]: I1203 07:35:23.921065 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l66pl" Dec 03 07:35:23 crc kubenswrapper[4818]: I1203 07:35:23.997889 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9kv6\" (UniqueName: \"kubernetes.io/projected/281a83c0-1e5e-4dac-b022-9904d0ac439e-kube-api-access-z9kv6\") pod \"281a83c0-1e5e-4dac-b022-9904d0ac439e\" (UID: \"281a83c0-1e5e-4dac-b022-9904d0ac439e\") " Dec 03 07:35:23 crc kubenswrapper[4818]: I1203 07:35:23.998100 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/281a83c0-1e5e-4dac-b022-9904d0ac439e-catalog-content\") pod \"281a83c0-1e5e-4dac-b022-9904d0ac439e\" (UID: \"281a83c0-1e5e-4dac-b022-9904d0ac439e\") " Dec 03 07:35:23 crc kubenswrapper[4818]: I1203 07:35:23.998230 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/281a83c0-1e5e-4dac-b022-9904d0ac439e-utilities\") pod \"281a83c0-1e5e-4dac-b022-9904d0ac439e\" (UID: \"281a83c0-1e5e-4dac-b022-9904d0ac439e\") " Dec 03 07:35:23 crc kubenswrapper[4818]: I1203 07:35:23.999528 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/281a83c0-1e5e-4dac-b022-9904d0ac439e-utilities" (OuterVolumeSpecName: "utilities") pod "281a83c0-1e5e-4dac-b022-9904d0ac439e" (UID: "281a83c0-1e5e-4dac-b022-9904d0ac439e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.003908 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/281a83c0-1e5e-4dac-b022-9904d0ac439e-kube-api-access-z9kv6" (OuterVolumeSpecName: "kube-api-access-z9kv6") pod "281a83c0-1e5e-4dac-b022-9904d0ac439e" (UID: "281a83c0-1e5e-4dac-b022-9904d0ac439e"). InnerVolumeSpecName "kube-api-access-z9kv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.045603 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/281a83c0-1e5e-4dac-b022-9904d0ac439e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "281a83c0-1e5e-4dac-b022-9904d0ac439e" (UID: "281a83c0-1e5e-4dac-b022-9904d0ac439e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.101096 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9kv6\" (UniqueName: \"kubernetes.io/projected/281a83c0-1e5e-4dac-b022-9904d0ac439e-kube-api-access-z9kv6\") on node \"crc\" DevicePath \"\"" Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.101135 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/281a83c0-1e5e-4dac-b022-9904d0ac439e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.101151 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/281a83c0-1e5e-4dac-b022-9904d0ac439e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.470248 4818 generic.go:334] "Generic (PLEG): container finished" podID="281a83c0-1e5e-4dac-b022-9904d0ac439e" containerID="f82ca608b54cfa94963278532a14affc4f8e2e407466a7aece9d87cc5fd9620d" exitCode=0 Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.470341 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l66pl" Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.470354 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l66pl" event={"ID":"281a83c0-1e5e-4dac-b022-9904d0ac439e","Type":"ContainerDied","Data":"f82ca608b54cfa94963278532a14affc4f8e2e407466a7aece9d87cc5fd9620d"} Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.470408 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l66pl" event={"ID":"281a83c0-1e5e-4dac-b022-9904d0ac439e","Type":"ContainerDied","Data":"9824faadd4c4d81b0a00b7aa33a31a623fc26de794d373c2e35468b219467eb6"} Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.470453 4818 scope.go:117] "RemoveContainer" containerID="f82ca608b54cfa94963278532a14affc4f8e2e407466a7aece9d87cc5fd9620d" Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.511662 4818 scope.go:117] "RemoveContainer" containerID="585d8d2cf9fdcf91e523dd19584c7c2dbba568e97fe2a5258256c2fc8af9901b" Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.526898 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l66pl"] Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.536242 4818 scope.go:117] "RemoveContainer" containerID="7882733c1d86342245ff2ce06a9b44a8acb4191acf6847ab421556af15c44f25" Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.538222 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l66pl"] Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.583211 4818 scope.go:117] "RemoveContainer" containerID="f82ca608b54cfa94963278532a14affc4f8e2e407466a7aece9d87cc5fd9620d" Dec 03 07:35:24 crc kubenswrapper[4818]: E1203 07:35:24.583879 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f82ca608b54cfa94963278532a14affc4f8e2e407466a7aece9d87cc5fd9620d\": container with ID starting with f82ca608b54cfa94963278532a14affc4f8e2e407466a7aece9d87cc5fd9620d not found: ID does not exist" containerID="f82ca608b54cfa94963278532a14affc4f8e2e407466a7aece9d87cc5fd9620d" Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.584064 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f82ca608b54cfa94963278532a14affc4f8e2e407466a7aece9d87cc5fd9620d"} err="failed to get container status \"f82ca608b54cfa94963278532a14affc4f8e2e407466a7aece9d87cc5fd9620d\": rpc error: code = NotFound desc = could not find container \"f82ca608b54cfa94963278532a14affc4f8e2e407466a7aece9d87cc5fd9620d\": container with ID starting with f82ca608b54cfa94963278532a14affc4f8e2e407466a7aece9d87cc5fd9620d not found: ID does not exist" Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.584193 4818 scope.go:117] "RemoveContainer" containerID="585d8d2cf9fdcf91e523dd19584c7c2dbba568e97fe2a5258256c2fc8af9901b" Dec 03 07:35:24 crc kubenswrapper[4818]: E1203 07:35:24.585235 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"585d8d2cf9fdcf91e523dd19584c7c2dbba568e97fe2a5258256c2fc8af9901b\": container with ID starting with 585d8d2cf9fdcf91e523dd19584c7c2dbba568e97fe2a5258256c2fc8af9901b not found: ID does not exist" containerID="585d8d2cf9fdcf91e523dd19584c7c2dbba568e97fe2a5258256c2fc8af9901b" Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.585284 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"585d8d2cf9fdcf91e523dd19584c7c2dbba568e97fe2a5258256c2fc8af9901b"} err="failed to get container status \"585d8d2cf9fdcf91e523dd19584c7c2dbba568e97fe2a5258256c2fc8af9901b\": rpc error: code = NotFound desc = could not find container \"585d8d2cf9fdcf91e523dd19584c7c2dbba568e97fe2a5258256c2fc8af9901b\": container with ID starting with 585d8d2cf9fdcf91e523dd19584c7c2dbba568e97fe2a5258256c2fc8af9901b not found: ID does not exist" Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.585317 4818 scope.go:117] "RemoveContainer" containerID="7882733c1d86342245ff2ce06a9b44a8acb4191acf6847ab421556af15c44f25" Dec 03 07:35:24 crc kubenswrapper[4818]: E1203 07:35:24.585745 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7882733c1d86342245ff2ce06a9b44a8acb4191acf6847ab421556af15c44f25\": container with ID starting with 7882733c1d86342245ff2ce06a9b44a8acb4191acf6847ab421556af15c44f25 not found: ID does not exist" containerID="7882733c1d86342245ff2ce06a9b44a8acb4191acf6847ab421556af15c44f25" Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.585774 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7882733c1d86342245ff2ce06a9b44a8acb4191acf6847ab421556af15c44f25"} err="failed to get container status \"7882733c1d86342245ff2ce06a9b44a8acb4191acf6847ab421556af15c44f25\": rpc error: code = NotFound desc = could not find container \"7882733c1d86342245ff2ce06a9b44a8acb4191acf6847ab421556af15c44f25\": container with ID starting with 7882733c1d86342245ff2ce06a9b44a8acb4191acf6847ab421556af15c44f25 not found: ID does not exist" Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.748179 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="281a83c0-1e5e-4dac-b022-9904d0ac439e" path="/var/lib/kubelet/pods/281a83c0-1e5e-4dac-b022-9904d0ac439e/volumes" Dec 03 07:35:24 crc kubenswrapper[4818]: I1203 07:35:24.853296 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8slz4" podUID="ef736fcb-502c-4a07-8b2f-cc0bc77297e9" containerName="registry-server" probeResult="failure" output=< Dec 03 07:35:24 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Dec 03 07:35:24 crc kubenswrapper[4818]: > Dec 03 07:35:25 crc kubenswrapper[4818]: I1203 07:35:25.228324 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xcqtm"] Dec 03 07:35:25 crc kubenswrapper[4818]: I1203 07:35:25.483037 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xcqtm" podUID="0f077309-37ed-4977-8638-86dacef46390" containerName="registry-server" containerID="cri-o://2e9f133b313e955e6848cac51386d402692adbae72697deec6d010220f6ec485" gracePeriod=2 Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.354333 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xcqtm" Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.450426 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrq6c\" (UniqueName: \"kubernetes.io/projected/0f077309-37ed-4977-8638-86dacef46390-kube-api-access-nrq6c\") pod \"0f077309-37ed-4977-8638-86dacef46390\" (UID: \"0f077309-37ed-4977-8638-86dacef46390\") " Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.450566 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f077309-37ed-4977-8638-86dacef46390-catalog-content\") pod \"0f077309-37ed-4977-8638-86dacef46390\" (UID: \"0f077309-37ed-4977-8638-86dacef46390\") " Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.450655 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f077309-37ed-4977-8638-86dacef46390-utilities\") pod \"0f077309-37ed-4977-8638-86dacef46390\" (UID: \"0f077309-37ed-4977-8638-86dacef46390\") " Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.451668 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f077309-37ed-4977-8638-86dacef46390-utilities" (OuterVolumeSpecName: "utilities") pod "0f077309-37ed-4977-8638-86dacef46390" (UID: "0f077309-37ed-4977-8638-86dacef46390"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.484959 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f077309-37ed-4977-8638-86dacef46390-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f077309-37ed-4977-8638-86dacef46390" (UID: "0f077309-37ed-4977-8638-86dacef46390"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.496332 4818 generic.go:334] "Generic (PLEG): container finished" podID="0f077309-37ed-4977-8638-86dacef46390" containerID="2e9f133b313e955e6848cac51386d402692adbae72697deec6d010220f6ec485" exitCode=0 Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.496446 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xcqtm" event={"ID":"0f077309-37ed-4977-8638-86dacef46390","Type":"ContainerDied","Data":"2e9f133b313e955e6848cac51386d402692adbae72697deec6d010220f6ec485"} Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.496616 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xcqtm" event={"ID":"0f077309-37ed-4977-8638-86dacef46390","Type":"ContainerDied","Data":"affb599f4cd96a888cca21dc0c45e8637a9bf3aed92b8937d0e47f69b9955c73"} Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.496708 4818 scope.go:117] "RemoveContainer" containerID="2e9f133b313e955e6848cac51386d402692adbae72697deec6d010220f6ec485" Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.496458 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xcqtm" Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.518144 4818 scope.go:117] "RemoveContainer" containerID="0489d5b2238e098100040a5c786647acd45cbc3a30242d4dcc58a4e5b3241104" Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.552375 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f077309-37ed-4977-8638-86dacef46390-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.552405 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f077309-37ed-4977-8638-86dacef46390-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.555969 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f077309-37ed-4977-8638-86dacef46390-kube-api-access-nrq6c" (OuterVolumeSpecName: "kube-api-access-nrq6c") pod "0f077309-37ed-4977-8638-86dacef46390" (UID: "0f077309-37ed-4977-8638-86dacef46390"). InnerVolumeSpecName "kube-api-access-nrq6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.567136 4818 scope.go:117] "RemoveContainer" containerID="be879c2f962530bd21ac89b79c90b00f60b27d12ee40d5914efedb644b3c2550" Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.583305 4818 scope.go:117] "RemoveContainer" containerID="2e9f133b313e955e6848cac51386d402692adbae72697deec6d010220f6ec485" Dec 03 07:35:26 crc kubenswrapper[4818]: E1203 07:35:26.583827 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e9f133b313e955e6848cac51386d402692adbae72697deec6d010220f6ec485\": container with ID starting with 2e9f133b313e955e6848cac51386d402692adbae72697deec6d010220f6ec485 not found: ID does not exist" containerID="2e9f133b313e955e6848cac51386d402692adbae72697deec6d010220f6ec485" Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.583872 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e9f133b313e955e6848cac51386d402692adbae72697deec6d010220f6ec485"} err="failed to get container status \"2e9f133b313e955e6848cac51386d402692adbae72697deec6d010220f6ec485\": rpc error: code = NotFound desc = could not find container \"2e9f133b313e955e6848cac51386d402692adbae72697deec6d010220f6ec485\": container with ID starting with 2e9f133b313e955e6848cac51386d402692adbae72697deec6d010220f6ec485 not found: ID does not exist" Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.583906 4818 scope.go:117] "RemoveContainer" containerID="0489d5b2238e098100040a5c786647acd45cbc3a30242d4dcc58a4e5b3241104" Dec 03 07:35:26 crc kubenswrapper[4818]: E1203 07:35:26.584338 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0489d5b2238e098100040a5c786647acd45cbc3a30242d4dcc58a4e5b3241104\": container with ID starting with 0489d5b2238e098100040a5c786647acd45cbc3a30242d4dcc58a4e5b3241104 not found: ID does not exist" containerID="0489d5b2238e098100040a5c786647acd45cbc3a30242d4dcc58a4e5b3241104" Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.584359 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0489d5b2238e098100040a5c786647acd45cbc3a30242d4dcc58a4e5b3241104"} err="failed to get container status \"0489d5b2238e098100040a5c786647acd45cbc3a30242d4dcc58a4e5b3241104\": rpc error: code = NotFound desc = could not find container \"0489d5b2238e098100040a5c786647acd45cbc3a30242d4dcc58a4e5b3241104\": container with ID starting with 0489d5b2238e098100040a5c786647acd45cbc3a30242d4dcc58a4e5b3241104 not found: ID does not exist" Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.584375 4818 scope.go:117] "RemoveContainer" containerID="be879c2f962530bd21ac89b79c90b00f60b27d12ee40d5914efedb644b3c2550" Dec 03 07:35:26 crc kubenswrapper[4818]: E1203 07:35:26.584838 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be879c2f962530bd21ac89b79c90b00f60b27d12ee40d5914efedb644b3c2550\": container with ID starting with be879c2f962530bd21ac89b79c90b00f60b27d12ee40d5914efedb644b3c2550 not found: ID does not exist" containerID="be879c2f962530bd21ac89b79c90b00f60b27d12ee40d5914efedb644b3c2550" Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.584878 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be879c2f962530bd21ac89b79c90b00f60b27d12ee40d5914efedb644b3c2550"} err="failed to get container status \"be879c2f962530bd21ac89b79c90b00f60b27d12ee40d5914efedb644b3c2550\": rpc error: code = NotFound desc = could not find container \"be879c2f962530bd21ac89b79c90b00f60b27d12ee40d5914efedb644b3c2550\": container with ID starting with be879c2f962530bd21ac89b79c90b00f60b27d12ee40d5914efedb644b3c2550 not found: ID does not exist" Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.654543 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrq6c\" (UniqueName: \"kubernetes.io/projected/0f077309-37ed-4977-8638-86dacef46390-kube-api-access-nrq6c\") on node \"crc\" DevicePath \"\"" Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.938828 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xcqtm"] Dec 03 07:35:26 crc kubenswrapper[4818]: I1203 07:35:26.947788 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xcqtm"] Dec 03 07:35:28 crc kubenswrapper[4818]: I1203 07:35:28.751089 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f077309-37ed-4977-8638-86dacef46390" path="/var/lib/kubelet/pods/0f077309-37ed-4977-8638-86dacef46390/volumes" Dec 03 07:35:33 crc kubenswrapper[4818]: I1203 07:35:33.861063 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8slz4" Dec 03 07:35:33 crc kubenswrapper[4818]: I1203 07:35:33.932176 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8slz4" Dec 03 07:35:34 crc kubenswrapper[4818]: I1203 07:35:34.099802 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8slz4"] Dec 03 07:35:35 crc kubenswrapper[4818]: I1203 07:35:35.580547 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8slz4" podUID="ef736fcb-502c-4a07-8b2f-cc0bc77297e9" containerName="registry-server" containerID="cri-o://eaa2db839379be9567bac92037002c79dec255f9f47316c13a95a7c5f5dc92ad" gracePeriod=2 Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.063981 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8slz4" Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.209865 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-catalog-content\") pod \"ef736fcb-502c-4a07-8b2f-cc0bc77297e9\" (UID: \"ef736fcb-502c-4a07-8b2f-cc0bc77297e9\") " Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.224576 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-utilities\") pod \"ef736fcb-502c-4a07-8b2f-cc0bc77297e9\" (UID: \"ef736fcb-502c-4a07-8b2f-cc0bc77297e9\") " Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.224730 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd54x\" (UniqueName: \"kubernetes.io/projected/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-kube-api-access-wd54x\") pod \"ef736fcb-502c-4a07-8b2f-cc0bc77297e9\" (UID: \"ef736fcb-502c-4a07-8b2f-cc0bc77297e9\") " Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.225347 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-utilities" (OuterVolumeSpecName: "utilities") pod "ef736fcb-502c-4a07-8b2f-cc0bc77297e9" (UID: "ef736fcb-502c-4a07-8b2f-cc0bc77297e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.225549 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.234856 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-kube-api-access-wd54x" (OuterVolumeSpecName: "kube-api-access-wd54x") pod "ef736fcb-502c-4a07-8b2f-cc0bc77297e9" (UID: "ef736fcb-502c-4a07-8b2f-cc0bc77297e9"). InnerVolumeSpecName "kube-api-access-wd54x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.326735 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ef736fcb-502c-4a07-8b2f-cc0bc77297e9" (UID: "ef736fcb-502c-4a07-8b2f-cc0bc77297e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.328133 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wd54x\" (UniqueName: \"kubernetes.io/projected/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-kube-api-access-wd54x\") on node \"crc\" DevicePath \"\"" Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.328157 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef736fcb-502c-4a07-8b2f-cc0bc77297e9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.590452 4818 generic.go:334] "Generic (PLEG): container finished" podID="ef736fcb-502c-4a07-8b2f-cc0bc77297e9" containerID="eaa2db839379be9567bac92037002c79dec255f9f47316c13a95a7c5f5dc92ad" exitCode=0 Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.590564 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8slz4" event={"ID":"ef736fcb-502c-4a07-8b2f-cc0bc77297e9","Type":"ContainerDied","Data":"eaa2db839379be9567bac92037002c79dec255f9f47316c13a95a7c5f5dc92ad"} Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.590905 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8slz4" event={"ID":"ef736fcb-502c-4a07-8b2f-cc0bc77297e9","Type":"ContainerDied","Data":"4caa594a06d9254ec48cafb41ed1560824e197b5bbd197793320f5d0cc0a33ed"} Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.590932 4818 scope.go:117] "RemoveContainer" containerID="eaa2db839379be9567bac92037002c79dec255f9f47316c13a95a7c5f5dc92ad" Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.590644 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8slz4" Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.614015 4818 scope.go:117] "RemoveContainer" containerID="cbe32f7471f49b87914db9ec53b4114168782924d378d14ccc2d211a4b1e8610" Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.645770 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8slz4"] Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.658636 4818 scope.go:117] "RemoveContainer" containerID="fefc67c0bfc45bf4f8dea70aaaba838229f346a629d064afdd99fe4b98d58ca4" Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.662086 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8slz4"] Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.698260 4818 scope.go:117] "RemoveContainer" containerID="eaa2db839379be9567bac92037002c79dec255f9f47316c13a95a7c5f5dc92ad" Dec 03 07:35:36 crc kubenswrapper[4818]: E1203 07:35:36.698809 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaa2db839379be9567bac92037002c79dec255f9f47316c13a95a7c5f5dc92ad\": container with ID starting with eaa2db839379be9567bac92037002c79dec255f9f47316c13a95a7c5f5dc92ad not found: ID does not exist" containerID="eaa2db839379be9567bac92037002c79dec255f9f47316c13a95a7c5f5dc92ad" Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.698881 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaa2db839379be9567bac92037002c79dec255f9f47316c13a95a7c5f5dc92ad"} err="failed to get container status \"eaa2db839379be9567bac92037002c79dec255f9f47316c13a95a7c5f5dc92ad\": rpc error: code = NotFound desc = could not find container \"eaa2db839379be9567bac92037002c79dec255f9f47316c13a95a7c5f5dc92ad\": container with ID starting with eaa2db839379be9567bac92037002c79dec255f9f47316c13a95a7c5f5dc92ad not found: ID does not exist" Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.698915 4818 scope.go:117] "RemoveContainer" containerID="cbe32f7471f49b87914db9ec53b4114168782924d378d14ccc2d211a4b1e8610" Dec 03 07:35:36 crc kubenswrapper[4818]: E1203 07:35:36.699252 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbe32f7471f49b87914db9ec53b4114168782924d378d14ccc2d211a4b1e8610\": container with ID starting with cbe32f7471f49b87914db9ec53b4114168782924d378d14ccc2d211a4b1e8610 not found: ID does not exist" containerID="cbe32f7471f49b87914db9ec53b4114168782924d378d14ccc2d211a4b1e8610" Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.699281 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbe32f7471f49b87914db9ec53b4114168782924d378d14ccc2d211a4b1e8610"} err="failed to get container status \"cbe32f7471f49b87914db9ec53b4114168782924d378d14ccc2d211a4b1e8610\": rpc error: code = NotFound desc = could not find container \"cbe32f7471f49b87914db9ec53b4114168782924d378d14ccc2d211a4b1e8610\": container with ID starting with cbe32f7471f49b87914db9ec53b4114168782924d378d14ccc2d211a4b1e8610 not found: ID does not exist" Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.699298 4818 scope.go:117] "RemoveContainer" containerID="fefc67c0bfc45bf4f8dea70aaaba838229f346a629d064afdd99fe4b98d58ca4" Dec 03 07:35:36 crc kubenswrapper[4818]: E1203 07:35:36.699625 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fefc67c0bfc45bf4f8dea70aaaba838229f346a629d064afdd99fe4b98d58ca4\": container with ID starting with fefc67c0bfc45bf4f8dea70aaaba838229f346a629d064afdd99fe4b98d58ca4 not found: ID does not exist" containerID="fefc67c0bfc45bf4f8dea70aaaba838229f346a629d064afdd99fe4b98d58ca4" Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.699660 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fefc67c0bfc45bf4f8dea70aaaba838229f346a629d064afdd99fe4b98d58ca4"} err="failed to get container status \"fefc67c0bfc45bf4f8dea70aaaba838229f346a629d064afdd99fe4b98d58ca4\": rpc error: code = NotFound desc = could not find container \"fefc67c0bfc45bf4f8dea70aaaba838229f346a629d064afdd99fe4b98d58ca4\": container with ID starting with fefc67c0bfc45bf4f8dea70aaaba838229f346a629d064afdd99fe4b98d58ca4 not found: ID does not exist" Dec 03 07:35:36 crc kubenswrapper[4818]: I1203 07:35:36.749180 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef736fcb-502c-4a07-8b2f-cc0bc77297e9" path="/var/lib/kubelet/pods/ef736fcb-502c-4a07-8b2f-cc0bc77297e9/volumes" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.790006 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-thj5g/must-gather-9qkqz"] Dec 03 07:35:38 crc kubenswrapper[4818]: E1203 07:35:38.791992 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f077309-37ed-4977-8638-86dacef46390" containerName="extract-content" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.792101 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f077309-37ed-4977-8638-86dacef46390" containerName="extract-content" Dec 03 07:35:38 crc kubenswrapper[4818]: E1203 07:35:38.792190 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="281a83c0-1e5e-4dac-b022-9904d0ac439e" containerName="extract-utilities" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.792282 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="281a83c0-1e5e-4dac-b022-9904d0ac439e" containerName="extract-utilities" Dec 03 07:35:38 crc kubenswrapper[4818]: E1203 07:35:38.792377 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="281a83c0-1e5e-4dac-b022-9904d0ac439e" containerName="extract-content" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.792453 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="281a83c0-1e5e-4dac-b022-9904d0ac439e" containerName="extract-content" Dec 03 07:35:38 crc kubenswrapper[4818]: E1203 07:35:38.792534 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="281a83c0-1e5e-4dac-b022-9904d0ac439e" containerName="registry-server" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.792620 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="281a83c0-1e5e-4dac-b022-9904d0ac439e" containerName="registry-server" Dec 03 07:35:38 crc kubenswrapper[4818]: E1203 07:35:38.792718 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef736fcb-502c-4a07-8b2f-cc0bc77297e9" containerName="extract-utilities" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.792802 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef736fcb-502c-4a07-8b2f-cc0bc77297e9" containerName="extract-utilities" Dec 03 07:35:38 crc kubenswrapper[4818]: E1203 07:35:38.792956 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef736fcb-502c-4a07-8b2f-cc0bc77297e9" containerName="registry-server" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.793048 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef736fcb-502c-4a07-8b2f-cc0bc77297e9" containerName="registry-server" Dec 03 07:35:38 crc kubenswrapper[4818]: E1203 07:35:38.793137 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f077309-37ed-4977-8638-86dacef46390" containerName="registry-server" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.793217 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f077309-37ed-4977-8638-86dacef46390" containerName="registry-server" Dec 03 07:35:38 crc kubenswrapper[4818]: E1203 07:35:38.793307 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef736fcb-502c-4a07-8b2f-cc0bc77297e9" containerName="extract-content" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.793405 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef736fcb-502c-4a07-8b2f-cc0bc77297e9" containerName="extract-content" Dec 03 07:35:38 crc kubenswrapper[4818]: E1203 07:35:38.793517 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f077309-37ed-4977-8638-86dacef46390" containerName="extract-utilities" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.793604 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f077309-37ed-4977-8638-86dacef46390" containerName="extract-utilities" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.794027 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef736fcb-502c-4a07-8b2f-cc0bc77297e9" containerName="registry-server" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.794149 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="281a83c0-1e5e-4dac-b022-9904d0ac439e" containerName="registry-server" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.794224 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f077309-37ed-4977-8638-86dacef46390" containerName="registry-server" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.795637 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-thj5g/must-gather-9qkqz" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.798124 4818 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-thj5g"/"default-dockercfg-6cbzg" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.802324 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-thj5g"/"kube-root-ca.crt" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.802357 4818 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-thj5g"/"openshift-service-ca.crt" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.811861 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-thj5g/must-gather-9qkqz"] Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.880028 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/af139d2b-a174-40c0-b8ae-d6db7c004db3-must-gather-output\") pod \"must-gather-9qkqz\" (UID: \"af139d2b-a174-40c0-b8ae-d6db7c004db3\") " pod="openshift-must-gather-thj5g/must-gather-9qkqz" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.880099 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtt4z\" (UniqueName: \"kubernetes.io/projected/af139d2b-a174-40c0-b8ae-d6db7c004db3-kube-api-access-gtt4z\") pod \"must-gather-9qkqz\" (UID: \"af139d2b-a174-40c0-b8ae-d6db7c004db3\") " pod="openshift-must-gather-thj5g/must-gather-9qkqz" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.981732 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtt4z\" (UniqueName: \"kubernetes.io/projected/af139d2b-a174-40c0-b8ae-d6db7c004db3-kube-api-access-gtt4z\") pod \"must-gather-9qkqz\" (UID: \"af139d2b-a174-40c0-b8ae-d6db7c004db3\") " pod="openshift-must-gather-thj5g/must-gather-9qkqz" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.981928 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/af139d2b-a174-40c0-b8ae-d6db7c004db3-must-gather-output\") pod \"must-gather-9qkqz\" (UID: \"af139d2b-a174-40c0-b8ae-d6db7c004db3\") " pod="openshift-must-gather-thj5g/must-gather-9qkqz" Dec 03 07:35:38 crc kubenswrapper[4818]: I1203 07:35:38.982298 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/af139d2b-a174-40c0-b8ae-d6db7c004db3-must-gather-output\") pod \"must-gather-9qkqz\" (UID: \"af139d2b-a174-40c0-b8ae-d6db7c004db3\") " pod="openshift-must-gather-thj5g/must-gather-9qkqz" Dec 03 07:35:39 crc kubenswrapper[4818]: I1203 07:35:39.013170 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtt4z\" (UniqueName: \"kubernetes.io/projected/af139d2b-a174-40c0-b8ae-d6db7c004db3-kube-api-access-gtt4z\") pod \"must-gather-9qkqz\" (UID: \"af139d2b-a174-40c0-b8ae-d6db7c004db3\") " pod="openshift-must-gather-thj5g/must-gather-9qkqz" Dec 03 07:35:39 crc kubenswrapper[4818]: I1203 07:35:39.114843 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-thj5g/must-gather-9qkqz" Dec 03 07:35:39 crc kubenswrapper[4818]: I1203 07:35:39.571964 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-thj5g/must-gather-9qkqz"] Dec 03 07:35:40 crc kubenswrapper[4818]: I1203 07:35:40.637005 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-thj5g/must-gather-9qkqz" event={"ID":"af139d2b-a174-40c0-b8ae-d6db7c004db3","Type":"ContainerStarted","Data":"c2d67076699bf0d1b5c53aebd9b2abfeca7a5bcbbbee3e1ff2306cc1d3f5aaa0"} Dec 03 07:35:40 crc kubenswrapper[4818]: I1203 07:35:40.637313 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-thj5g/must-gather-9qkqz" event={"ID":"af139d2b-a174-40c0-b8ae-d6db7c004db3","Type":"ContainerStarted","Data":"1d41ea4397ee8fe6b1080387c5816db86b4092454b715fc376677ac73dac9e7b"} Dec 03 07:35:41 crc kubenswrapper[4818]: I1203 07:35:41.651849 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-thj5g/must-gather-9qkqz" event={"ID":"af139d2b-a174-40c0-b8ae-d6db7c004db3","Type":"ContainerStarted","Data":"1d3e56afafdfed16684e2932255bdbb3c45bb35aba8a20339aaf2d3a2d2ff0b1"} Dec 03 07:35:41 crc kubenswrapper[4818]: I1203 07:35:41.677004 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-thj5g/must-gather-9qkqz" podStartSLOduration=3.676986018 podStartE2EDuration="3.676986018s" podCreationTimestamp="2025-12-03 07:35:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:35:41.670145598 +0000 UTC m=+4099.361754350" watchObservedRunningTime="2025-12-03 07:35:41.676986018 +0000 UTC m=+4099.368594770" Dec 03 07:35:43 crc kubenswrapper[4818]: I1203 07:35:43.302585 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:35:43 crc kubenswrapper[4818]: I1203 07:35:43.303039 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:35:43 crc kubenswrapper[4818]: I1203 07:35:43.919634 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-thj5g/crc-debug-gmcxd"] Dec 03 07:35:43 crc kubenswrapper[4818]: I1203 07:35:43.921230 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-thj5g/crc-debug-gmcxd" Dec 03 07:35:44 crc kubenswrapper[4818]: I1203 07:35:44.085554 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f693ec9-d8d6-460d-ac3a-77287492cab7-host\") pod \"crc-debug-gmcxd\" (UID: \"9f693ec9-d8d6-460d-ac3a-77287492cab7\") " pod="openshift-must-gather-thj5g/crc-debug-gmcxd" Dec 03 07:35:44 crc kubenswrapper[4818]: I1203 07:35:44.085981 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbg26\" (UniqueName: \"kubernetes.io/projected/9f693ec9-d8d6-460d-ac3a-77287492cab7-kube-api-access-gbg26\") pod \"crc-debug-gmcxd\" (UID: \"9f693ec9-d8d6-460d-ac3a-77287492cab7\") " pod="openshift-must-gather-thj5g/crc-debug-gmcxd" Dec 03 07:35:44 crc kubenswrapper[4818]: I1203 07:35:44.188093 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f693ec9-d8d6-460d-ac3a-77287492cab7-host\") pod \"crc-debug-gmcxd\" (UID: \"9f693ec9-d8d6-460d-ac3a-77287492cab7\") " pod="openshift-must-gather-thj5g/crc-debug-gmcxd" Dec 03 07:35:44 crc kubenswrapper[4818]: I1203 07:35:44.188162 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbg26\" (UniqueName: \"kubernetes.io/projected/9f693ec9-d8d6-460d-ac3a-77287492cab7-kube-api-access-gbg26\") pod \"crc-debug-gmcxd\" (UID: \"9f693ec9-d8d6-460d-ac3a-77287492cab7\") " pod="openshift-must-gather-thj5g/crc-debug-gmcxd" Dec 03 07:35:44 crc kubenswrapper[4818]: I1203 07:35:44.188237 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f693ec9-d8d6-460d-ac3a-77287492cab7-host\") pod \"crc-debug-gmcxd\" (UID: \"9f693ec9-d8d6-460d-ac3a-77287492cab7\") " pod="openshift-must-gather-thj5g/crc-debug-gmcxd" Dec 03 07:35:44 crc kubenswrapper[4818]: I1203 07:35:44.208095 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbg26\" (UniqueName: \"kubernetes.io/projected/9f693ec9-d8d6-460d-ac3a-77287492cab7-kube-api-access-gbg26\") pod \"crc-debug-gmcxd\" (UID: \"9f693ec9-d8d6-460d-ac3a-77287492cab7\") " pod="openshift-must-gather-thj5g/crc-debug-gmcxd" Dec 03 07:35:44 crc kubenswrapper[4818]: I1203 07:35:44.252927 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-thj5g/crc-debug-gmcxd" Dec 03 07:35:44 crc kubenswrapper[4818]: W1203 07:35:44.321002 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f693ec9_d8d6_460d_ac3a_77287492cab7.slice/crio-d71c1d827b3609dcb3977db63d0b6858631197da604e2eaf6f6255087e4df8ee WatchSource:0}: Error finding container d71c1d827b3609dcb3977db63d0b6858631197da604e2eaf6f6255087e4df8ee: Status 404 returned error can't find the container with id d71c1d827b3609dcb3977db63d0b6858631197da604e2eaf6f6255087e4df8ee Dec 03 07:35:44 crc kubenswrapper[4818]: I1203 07:35:44.677212 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-thj5g/crc-debug-gmcxd" event={"ID":"9f693ec9-d8d6-460d-ac3a-77287492cab7","Type":"ContainerStarted","Data":"b04be86f5dc7cf57c0c8e45240d48e298868a3f19d8d2c66f49c32285ecccd0f"} Dec 03 07:35:44 crc kubenswrapper[4818]: I1203 07:35:44.677445 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-thj5g/crc-debug-gmcxd" event={"ID":"9f693ec9-d8d6-460d-ac3a-77287492cab7","Type":"ContainerStarted","Data":"d71c1d827b3609dcb3977db63d0b6858631197da604e2eaf6f6255087e4df8ee"} Dec 03 07:35:44 crc kubenswrapper[4818]: I1203 07:35:44.695721 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-thj5g/crc-debug-gmcxd" podStartSLOduration=1.695700643 podStartE2EDuration="1.695700643s" podCreationTimestamp="2025-12-03 07:35:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 07:35:44.691693263 +0000 UTC m=+4102.383302015" watchObservedRunningTime="2025-12-03 07:35:44.695700643 +0000 UTC m=+4102.387309395" Dec 03 07:36:13 crc kubenswrapper[4818]: I1203 07:36:13.303023 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:36:13 crc kubenswrapper[4818]: I1203 07:36:13.303616 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:36:13 crc kubenswrapper[4818]: I1203 07:36:13.303675 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 07:36:13 crc kubenswrapper[4818]: I1203 07:36:13.304457 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3696ecd804b3cc54d22f8baf9c3f030c9a0933db13e3f5ad4cac7c6ea9cfefef"} pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:36:13 crc kubenswrapper[4818]: I1203 07:36:13.304522 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" containerID="cri-o://3696ecd804b3cc54d22f8baf9c3f030c9a0933db13e3f5ad4cac7c6ea9cfefef" gracePeriod=600 Dec 03 07:36:13 crc kubenswrapper[4818]: I1203 07:36:13.924832 4818 generic.go:334] "Generic (PLEG): container finished" podID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerID="3696ecd804b3cc54d22f8baf9c3f030c9a0933db13e3f5ad4cac7c6ea9cfefef" exitCode=0 Dec 03 07:36:13 crc kubenswrapper[4818]: I1203 07:36:13.925402 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerDied","Data":"3696ecd804b3cc54d22f8baf9c3f030c9a0933db13e3f5ad4cac7c6ea9cfefef"} Dec 03 07:36:13 crc kubenswrapper[4818]: I1203 07:36:13.925515 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595"} Dec 03 07:36:13 crc kubenswrapper[4818]: I1203 07:36:13.925604 4818 scope.go:117] "RemoveContainer" containerID="e39348ff934db2b4212deb02ca101ca3accc2d652504def86ff1c37680e5d020" Dec 03 07:36:18 crc kubenswrapper[4818]: I1203 07:36:18.988493 4818 generic.go:334] "Generic (PLEG): container finished" podID="9f693ec9-d8d6-460d-ac3a-77287492cab7" containerID="b04be86f5dc7cf57c0c8e45240d48e298868a3f19d8d2c66f49c32285ecccd0f" exitCode=0 Dec 03 07:36:18 crc kubenswrapper[4818]: I1203 07:36:18.988589 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-thj5g/crc-debug-gmcxd" event={"ID":"9f693ec9-d8d6-460d-ac3a-77287492cab7","Type":"ContainerDied","Data":"b04be86f5dc7cf57c0c8e45240d48e298868a3f19d8d2c66f49c32285ecccd0f"} Dec 03 07:36:20 crc kubenswrapper[4818]: I1203 07:36:20.480545 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-thj5g/crc-debug-gmcxd" Dec 03 07:36:20 crc kubenswrapper[4818]: I1203 07:36:20.517952 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-thj5g/crc-debug-gmcxd"] Dec 03 07:36:20 crc kubenswrapper[4818]: I1203 07:36:20.525688 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-thj5g/crc-debug-gmcxd"] Dec 03 07:36:20 crc kubenswrapper[4818]: I1203 07:36:20.582107 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbg26\" (UniqueName: \"kubernetes.io/projected/9f693ec9-d8d6-460d-ac3a-77287492cab7-kube-api-access-gbg26\") pod \"9f693ec9-d8d6-460d-ac3a-77287492cab7\" (UID: \"9f693ec9-d8d6-460d-ac3a-77287492cab7\") " Dec 03 07:36:20 crc kubenswrapper[4818]: I1203 07:36:20.582164 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f693ec9-d8d6-460d-ac3a-77287492cab7-host\") pod \"9f693ec9-d8d6-460d-ac3a-77287492cab7\" (UID: \"9f693ec9-d8d6-460d-ac3a-77287492cab7\") " Dec 03 07:36:20 crc kubenswrapper[4818]: I1203 07:36:20.582527 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f693ec9-d8d6-460d-ac3a-77287492cab7-host" (OuterVolumeSpecName: "host") pod "9f693ec9-d8d6-460d-ac3a-77287492cab7" (UID: "9f693ec9-d8d6-460d-ac3a-77287492cab7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:36:20 crc kubenswrapper[4818]: I1203 07:36:20.587414 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f693ec9-d8d6-460d-ac3a-77287492cab7-kube-api-access-gbg26" (OuterVolumeSpecName: "kube-api-access-gbg26") pod "9f693ec9-d8d6-460d-ac3a-77287492cab7" (UID: "9f693ec9-d8d6-460d-ac3a-77287492cab7"). InnerVolumeSpecName "kube-api-access-gbg26". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:36:20 crc kubenswrapper[4818]: I1203 07:36:20.684187 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbg26\" (UniqueName: \"kubernetes.io/projected/9f693ec9-d8d6-460d-ac3a-77287492cab7-kube-api-access-gbg26\") on node \"crc\" DevicePath \"\"" Dec 03 07:36:20 crc kubenswrapper[4818]: I1203 07:36:20.684224 4818 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9f693ec9-d8d6-460d-ac3a-77287492cab7-host\") on node \"crc\" DevicePath \"\"" Dec 03 07:36:20 crc kubenswrapper[4818]: I1203 07:36:20.748709 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f693ec9-d8d6-460d-ac3a-77287492cab7" path="/var/lib/kubelet/pods/9f693ec9-d8d6-460d-ac3a-77287492cab7/volumes" Dec 03 07:36:21 crc kubenswrapper[4818]: I1203 07:36:21.012849 4818 scope.go:117] "RemoveContainer" containerID="b04be86f5dc7cf57c0c8e45240d48e298868a3f19d8d2c66f49c32285ecccd0f" Dec 03 07:36:21 crc kubenswrapper[4818]: I1203 07:36:21.012888 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-thj5g/crc-debug-gmcxd" Dec 03 07:36:21 crc kubenswrapper[4818]: I1203 07:36:21.723672 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-thj5g/crc-debug-r279r"] Dec 03 07:36:21 crc kubenswrapper[4818]: E1203 07:36:21.724138 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f693ec9-d8d6-460d-ac3a-77287492cab7" containerName="container-00" Dec 03 07:36:21 crc kubenswrapper[4818]: I1203 07:36:21.724152 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f693ec9-d8d6-460d-ac3a-77287492cab7" containerName="container-00" Dec 03 07:36:21 crc kubenswrapper[4818]: I1203 07:36:21.724333 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f693ec9-d8d6-460d-ac3a-77287492cab7" containerName="container-00" Dec 03 07:36:21 crc kubenswrapper[4818]: I1203 07:36:21.725081 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-thj5g/crc-debug-r279r" Dec 03 07:36:21 crc kubenswrapper[4818]: I1203 07:36:21.908920 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e91c05d1-885e-4dbf-bbfd-643ea9bc18a1-host\") pod \"crc-debug-r279r\" (UID: \"e91c05d1-885e-4dbf-bbfd-643ea9bc18a1\") " pod="openshift-must-gather-thj5g/crc-debug-r279r" Dec 03 07:36:21 crc kubenswrapper[4818]: I1203 07:36:21.910673 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rz98\" (UniqueName: \"kubernetes.io/projected/e91c05d1-885e-4dbf-bbfd-643ea9bc18a1-kube-api-access-9rz98\") pod \"crc-debug-r279r\" (UID: \"e91c05d1-885e-4dbf-bbfd-643ea9bc18a1\") " pod="openshift-must-gather-thj5g/crc-debug-r279r" Dec 03 07:36:22 crc kubenswrapper[4818]: I1203 07:36:22.012729 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e91c05d1-885e-4dbf-bbfd-643ea9bc18a1-host\") pod \"crc-debug-r279r\" (UID: \"e91c05d1-885e-4dbf-bbfd-643ea9bc18a1\") " pod="openshift-must-gather-thj5g/crc-debug-r279r" Dec 03 07:36:22 crc kubenswrapper[4818]: I1203 07:36:22.012847 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rz98\" (UniqueName: \"kubernetes.io/projected/e91c05d1-885e-4dbf-bbfd-643ea9bc18a1-kube-api-access-9rz98\") pod \"crc-debug-r279r\" (UID: \"e91c05d1-885e-4dbf-bbfd-643ea9bc18a1\") " pod="openshift-must-gather-thj5g/crc-debug-r279r" Dec 03 07:36:22 crc kubenswrapper[4818]: I1203 07:36:22.012930 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e91c05d1-885e-4dbf-bbfd-643ea9bc18a1-host\") pod \"crc-debug-r279r\" (UID: \"e91c05d1-885e-4dbf-bbfd-643ea9bc18a1\") " pod="openshift-must-gather-thj5g/crc-debug-r279r" Dec 03 07:36:22 crc kubenswrapper[4818]: I1203 07:36:22.043332 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rz98\" (UniqueName: \"kubernetes.io/projected/e91c05d1-885e-4dbf-bbfd-643ea9bc18a1-kube-api-access-9rz98\") pod \"crc-debug-r279r\" (UID: \"e91c05d1-885e-4dbf-bbfd-643ea9bc18a1\") " pod="openshift-must-gather-thj5g/crc-debug-r279r" Dec 03 07:36:22 crc kubenswrapper[4818]: I1203 07:36:22.342079 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-thj5g/crc-debug-r279r" Dec 03 07:36:23 crc kubenswrapper[4818]: I1203 07:36:23.031962 4818 generic.go:334] "Generic (PLEG): container finished" podID="e91c05d1-885e-4dbf-bbfd-643ea9bc18a1" containerID="0d03852c4a6a0cd0a137036e8ba184e9ce05767d1d00b9a26397687fce19e8a7" exitCode=0 Dec 03 07:36:23 crc kubenswrapper[4818]: I1203 07:36:23.032050 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-thj5g/crc-debug-r279r" event={"ID":"e91c05d1-885e-4dbf-bbfd-643ea9bc18a1","Type":"ContainerDied","Data":"0d03852c4a6a0cd0a137036e8ba184e9ce05767d1d00b9a26397687fce19e8a7"} Dec 03 07:36:23 crc kubenswrapper[4818]: I1203 07:36:23.032577 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-thj5g/crc-debug-r279r" event={"ID":"e91c05d1-885e-4dbf-bbfd-643ea9bc18a1","Type":"ContainerStarted","Data":"15e36f67965ecac519643859b549a0f5d7f04f22e371aae5e442c1f45cb70bf3"} Dec 03 07:36:23 crc kubenswrapper[4818]: I1203 07:36:23.478245 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-thj5g/crc-debug-r279r"] Dec 03 07:36:23 crc kubenswrapper[4818]: I1203 07:36:23.486752 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-thj5g/crc-debug-r279r"] Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.145454 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-thj5g/crc-debug-r279r" Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.256413 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rz98\" (UniqueName: \"kubernetes.io/projected/e91c05d1-885e-4dbf-bbfd-643ea9bc18a1-kube-api-access-9rz98\") pod \"e91c05d1-885e-4dbf-bbfd-643ea9bc18a1\" (UID: \"e91c05d1-885e-4dbf-bbfd-643ea9bc18a1\") " Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.256634 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e91c05d1-885e-4dbf-bbfd-643ea9bc18a1-host\") pod \"e91c05d1-885e-4dbf-bbfd-643ea9bc18a1\" (UID: \"e91c05d1-885e-4dbf-bbfd-643ea9bc18a1\") " Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.257209 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e91c05d1-885e-4dbf-bbfd-643ea9bc18a1-host" (OuterVolumeSpecName: "host") pod "e91c05d1-885e-4dbf-bbfd-643ea9bc18a1" (UID: "e91c05d1-885e-4dbf-bbfd-643ea9bc18a1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.268528 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e91c05d1-885e-4dbf-bbfd-643ea9bc18a1-kube-api-access-9rz98" (OuterVolumeSpecName: "kube-api-access-9rz98") pod "e91c05d1-885e-4dbf-bbfd-643ea9bc18a1" (UID: "e91c05d1-885e-4dbf-bbfd-643ea9bc18a1"). InnerVolumeSpecName "kube-api-access-9rz98". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.358834 4818 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e91c05d1-885e-4dbf-bbfd-643ea9bc18a1-host\") on node \"crc\" DevicePath \"\"" Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.358864 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rz98\" (UniqueName: \"kubernetes.io/projected/e91c05d1-885e-4dbf-bbfd-643ea9bc18a1-kube-api-access-9rz98\") on node \"crc\" DevicePath \"\"" Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.648420 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-thj5g/crc-debug-zk9p8"] Dec 03 07:36:24 crc kubenswrapper[4818]: E1203 07:36:24.648888 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e91c05d1-885e-4dbf-bbfd-643ea9bc18a1" containerName="container-00" Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.648914 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="e91c05d1-885e-4dbf-bbfd-643ea9bc18a1" containerName="container-00" Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.649241 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="e91c05d1-885e-4dbf-bbfd-643ea9bc18a1" containerName="container-00" Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.650109 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-thj5g/crc-debug-zk9p8" Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.750152 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e91c05d1-885e-4dbf-bbfd-643ea9bc18a1" path="/var/lib/kubelet/pods/e91c05d1-885e-4dbf-bbfd-643ea9bc18a1/volumes" Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.777587 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b629x\" (UniqueName: \"kubernetes.io/projected/4acb60dd-6f98-4c73-b65b-31393936d2f9-kube-api-access-b629x\") pod \"crc-debug-zk9p8\" (UID: \"4acb60dd-6f98-4c73-b65b-31393936d2f9\") " pod="openshift-must-gather-thj5g/crc-debug-zk9p8" Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.777686 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4acb60dd-6f98-4c73-b65b-31393936d2f9-host\") pod \"crc-debug-zk9p8\" (UID: \"4acb60dd-6f98-4c73-b65b-31393936d2f9\") " pod="openshift-must-gather-thj5g/crc-debug-zk9p8" Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.879333 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4acb60dd-6f98-4c73-b65b-31393936d2f9-host\") pod \"crc-debug-zk9p8\" (UID: \"4acb60dd-6f98-4c73-b65b-31393936d2f9\") " pod="openshift-must-gather-thj5g/crc-debug-zk9p8" Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.879523 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4acb60dd-6f98-4c73-b65b-31393936d2f9-host\") pod \"crc-debug-zk9p8\" (UID: \"4acb60dd-6f98-4c73-b65b-31393936d2f9\") " pod="openshift-must-gather-thj5g/crc-debug-zk9p8" Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.879938 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b629x\" (UniqueName: \"kubernetes.io/projected/4acb60dd-6f98-4c73-b65b-31393936d2f9-kube-api-access-b629x\") pod \"crc-debug-zk9p8\" (UID: \"4acb60dd-6f98-4c73-b65b-31393936d2f9\") " pod="openshift-must-gather-thj5g/crc-debug-zk9p8" Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.905793 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b629x\" (UniqueName: \"kubernetes.io/projected/4acb60dd-6f98-4c73-b65b-31393936d2f9-kube-api-access-b629x\") pod \"crc-debug-zk9p8\" (UID: \"4acb60dd-6f98-4c73-b65b-31393936d2f9\") " pod="openshift-must-gather-thj5g/crc-debug-zk9p8" Dec 03 07:36:24 crc kubenswrapper[4818]: I1203 07:36:24.979413 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-thj5g/crc-debug-zk9p8" Dec 03 07:36:25 crc kubenswrapper[4818]: W1203 07:36:25.033772 4818 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4acb60dd_6f98_4c73_b65b_31393936d2f9.slice/crio-3fb78689e8a0645ede556f06eb4002b4fea9114fba7b128cd8da92e35b78cb09 WatchSource:0}: Error finding container 3fb78689e8a0645ede556f06eb4002b4fea9114fba7b128cd8da92e35b78cb09: Status 404 returned error can't find the container with id 3fb78689e8a0645ede556f06eb4002b4fea9114fba7b128cd8da92e35b78cb09 Dec 03 07:36:25 crc kubenswrapper[4818]: I1203 07:36:25.051756 4818 scope.go:117] "RemoveContainer" containerID="0d03852c4a6a0cd0a137036e8ba184e9ce05767d1d00b9a26397687fce19e8a7" Dec 03 07:36:25 crc kubenswrapper[4818]: I1203 07:36:25.051775 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-thj5g/crc-debug-r279r" Dec 03 07:36:25 crc kubenswrapper[4818]: I1203 07:36:25.055282 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-thj5g/crc-debug-zk9p8" event={"ID":"4acb60dd-6f98-4c73-b65b-31393936d2f9","Type":"ContainerStarted","Data":"3fb78689e8a0645ede556f06eb4002b4fea9114fba7b128cd8da92e35b78cb09"} Dec 03 07:36:26 crc kubenswrapper[4818]: I1203 07:36:26.069681 4818 generic.go:334] "Generic (PLEG): container finished" podID="4acb60dd-6f98-4c73-b65b-31393936d2f9" containerID="f5b8ce276ca40d34493ab2af4fc3572762aa7e7dbf573c94cd5d886f5602b879" exitCode=0 Dec 03 07:36:26 crc kubenswrapper[4818]: I1203 07:36:26.070032 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-thj5g/crc-debug-zk9p8" event={"ID":"4acb60dd-6f98-4c73-b65b-31393936d2f9","Type":"ContainerDied","Data":"f5b8ce276ca40d34493ab2af4fc3572762aa7e7dbf573c94cd5d886f5602b879"} Dec 03 07:36:26 crc kubenswrapper[4818]: I1203 07:36:26.114435 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-thj5g/crc-debug-zk9p8"] Dec 03 07:36:26 crc kubenswrapper[4818]: I1203 07:36:26.126753 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-thj5g/crc-debug-zk9p8"] Dec 03 07:36:27 crc kubenswrapper[4818]: I1203 07:36:27.190321 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-thj5g/crc-debug-zk9p8" Dec 03 07:36:27 crc kubenswrapper[4818]: I1203 07:36:27.327986 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b629x\" (UniqueName: \"kubernetes.io/projected/4acb60dd-6f98-4c73-b65b-31393936d2f9-kube-api-access-b629x\") pod \"4acb60dd-6f98-4c73-b65b-31393936d2f9\" (UID: \"4acb60dd-6f98-4c73-b65b-31393936d2f9\") " Dec 03 07:36:27 crc kubenswrapper[4818]: I1203 07:36:27.328218 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4acb60dd-6f98-4c73-b65b-31393936d2f9-host\") pod \"4acb60dd-6f98-4c73-b65b-31393936d2f9\" (UID: \"4acb60dd-6f98-4c73-b65b-31393936d2f9\") " Dec 03 07:36:27 crc kubenswrapper[4818]: I1203 07:36:27.328838 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4acb60dd-6f98-4c73-b65b-31393936d2f9-host" (OuterVolumeSpecName: "host") pod "4acb60dd-6f98-4c73-b65b-31393936d2f9" (UID: "4acb60dd-6f98-4c73-b65b-31393936d2f9"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 07:36:27 crc kubenswrapper[4818]: I1203 07:36:27.336098 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4acb60dd-6f98-4c73-b65b-31393936d2f9-kube-api-access-b629x" (OuterVolumeSpecName: "kube-api-access-b629x") pod "4acb60dd-6f98-4c73-b65b-31393936d2f9" (UID: "4acb60dd-6f98-4c73-b65b-31393936d2f9"). InnerVolumeSpecName "kube-api-access-b629x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:36:27 crc kubenswrapper[4818]: I1203 07:36:27.431183 4818 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4acb60dd-6f98-4c73-b65b-31393936d2f9-host\") on node \"crc\" DevicePath \"\"" Dec 03 07:36:27 crc kubenswrapper[4818]: I1203 07:36:27.431555 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b629x\" (UniqueName: \"kubernetes.io/projected/4acb60dd-6f98-4c73-b65b-31393936d2f9-kube-api-access-b629x\") on node \"crc\" DevicePath \"\"" Dec 03 07:36:28 crc kubenswrapper[4818]: I1203 07:36:28.090805 4818 scope.go:117] "RemoveContainer" containerID="f5b8ce276ca40d34493ab2af4fc3572762aa7e7dbf573c94cd5d886f5602b879" Dec 03 07:36:28 crc kubenswrapper[4818]: I1203 07:36:28.090860 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-thj5g/crc-debug-zk9p8" Dec 03 07:36:28 crc kubenswrapper[4818]: I1203 07:36:28.747855 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4acb60dd-6f98-4c73-b65b-31393936d2f9" path="/var/lib/kubelet/pods/4acb60dd-6f98-4c73-b65b-31393936d2f9/volumes" Dec 03 07:36:52 crc kubenswrapper[4818]: I1203 07:36:52.356808 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7454b897fd-qzp7b_5e283b07-3849-4796-9b82-84e5ca0bf431/barbican-api/0.log" Dec 03 07:36:52 crc kubenswrapper[4818]: I1203 07:36:52.371273 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7454b897fd-qzp7b_5e283b07-3849-4796-9b82-84e5ca0bf431/barbican-api-log/0.log" Dec 03 07:36:52 crc kubenswrapper[4818]: I1203 07:36:52.577859 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7b4848f678-kfvpp_6d8d8f21-1f1f-4431-8079-ffad453ff3f1/barbican-keystone-listener/0.log" Dec 03 07:36:52 crc kubenswrapper[4818]: I1203 07:36:52.594958 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7b4848f678-kfvpp_6d8d8f21-1f1f-4431-8079-ffad453ff3f1/barbican-keystone-listener-log/0.log" Dec 03 07:36:52 crc kubenswrapper[4818]: I1203 07:36:52.613781 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-554df66f87-jscgb_d8e58804-c7d8-48ce-9db3-6c3ab56f1c13/barbican-worker/0.log" Dec 03 07:36:52 crc kubenswrapper[4818]: I1203 07:36:52.767113 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-554df66f87-jscgb_d8e58804-c7d8-48ce-9db3-6c3ab56f1c13/barbican-worker-log/0.log" Dec 03 07:36:52 crc kubenswrapper[4818]: I1203 07:36:52.804011 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-jph5g_2b2dc637-0ece-412a-b1e2-413dff6a4f08/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:36:52 crc kubenswrapper[4818]: I1203 07:36:52.978481 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_631b9437-fe40-43b7-b8ad-34b4cad6ffa1/ceilometer-central-agent/0.log" Dec 03 07:36:53 crc kubenswrapper[4818]: I1203 07:36:53.002720 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_631b9437-fe40-43b7-b8ad-34b4cad6ffa1/proxy-httpd/0.log" Dec 03 07:36:53 crc kubenswrapper[4818]: I1203 07:36:53.007939 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_631b9437-fe40-43b7-b8ad-34b4cad6ffa1/ceilometer-notification-agent/0.log" Dec 03 07:36:53 crc kubenswrapper[4818]: I1203 07:36:53.055692 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_631b9437-fe40-43b7-b8ad-34b4cad6ffa1/sg-core/0.log" Dec 03 07:36:53 crc kubenswrapper[4818]: I1203 07:36:53.181050 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_462028f4-61c0-4862-8707-3981513c3b76/cinder-api-log/0.log" Dec 03 07:36:53 crc kubenswrapper[4818]: I1203 07:36:53.219894 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_462028f4-61c0-4862-8707-3981513c3b76/cinder-api/0.log" Dec 03 07:36:53 crc kubenswrapper[4818]: I1203 07:36:53.473681 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5ad1caca-cc5b-4625-8776-2fdc829594a4/probe/0.log" Dec 03 07:36:53 crc kubenswrapper[4818]: I1203 07:36:53.499042 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-hzwrn_71c50082-f6aa-4099-be63-2b33234f0636/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:36:53 crc kubenswrapper[4818]: I1203 07:36:53.502486 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5ad1caca-cc5b-4625-8776-2fdc829594a4/cinder-scheduler/0.log" Dec 03 07:36:53 crc kubenswrapper[4818]: I1203 07:36:53.649474 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-2c8t9_2ef6c961-25b1-4f6f-a98b-65d79787e9e1/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:36:53 crc kubenswrapper[4818]: I1203 07:36:53.731354 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-gbkbl_19b91d65-7f62-47e7-b12c-33135811b76b/init/0.log" Dec 03 07:36:53 crc kubenswrapper[4818]: I1203 07:36:53.889185 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-gbkbl_19b91d65-7f62-47e7-b12c-33135811b76b/init/0.log" Dec 03 07:36:53 crc kubenswrapper[4818]: I1203 07:36:53.961805 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-gbkbl_19b91d65-7f62-47e7-b12c-33135811b76b/dnsmasq-dns/0.log" Dec 03 07:36:53 crc kubenswrapper[4818]: I1203 07:36:53.992023 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-fsmjq_86dc24e9-831f-4731-a612-98009562639c/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:36:54 crc kubenswrapper[4818]: I1203 07:36:54.347474 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0c46a9a7-6f86-40b8-8957-5d0f11d11262/glance-httpd/0.log" Dec 03 07:36:54 crc kubenswrapper[4818]: I1203 07:36:54.393013 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0c46a9a7-6f86-40b8-8957-5d0f11d11262/glance-log/0.log" Dec 03 07:36:54 crc kubenswrapper[4818]: I1203 07:36:54.592491 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_22a6b677-2df1-41ed-a420-dbc5ce60afc5/glance-log/0.log" Dec 03 07:36:54 crc kubenswrapper[4818]: I1203 07:36:54.604348 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_22a6b677-2df1-41ed-a420-dbc5ce60afc5/glance-httpd/0.log" Dec 03 07:36:54 crc kubenswrapper[4818]: I1203 07:36:54.839023 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-z89w5_1808af84-64d0-4b2a-a5bf-32653bb3c948/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:36:54 crc kubenswrapper[4818]: I1203 07:36:54.855747 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-68b68c8c84-qt25s_23c89695-46b0-49d8-ab09-eb002d3cb94e/horizon/0.log" Dec 03 07:36:55 crc kubenswrapper[4818]: I1203 07:36:55.104042 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-pdfz4_9568e2c6-4f91-49ce-80e6-0081b7bef2c5/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:36:55 crc kubenswrapper[4818]: I1203 07:36:55.161532 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-68b68c8c84-qt25s_23c89695-46b0-49d8-ab09-eb002d3cb94e/horizon-log/0.log" Dec 03 07:36:55 crc kubenswrapper[4818]: I1203 07:36:55.317157 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29412421-4n5xh_658c06c1-51a3-467d-8de0-2205f7c22cbc/keystone-cron/0.log" Dec 03 07:36:55 crc kubenswrapper[4818]: I1203 07:36:55.497900 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-75b5b56458-pbq6h_29098cb1-d622-4ee2-ad26-f8d51eec41f4/keystone-api/0.log" Dec 03 07:36:55 crc kubenswrapper[4818]: I1203 07:36:55.543865 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_0d105427-cef6-4ec6-8997-08f8a94e5b56/kube-state-metrics/0.log" Dec 03 07:36:55 crc kubenswrapper[4818]: I1203 07:36:55.647925 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-mtzwd_e5796328-5b74-4f5a-939b-403485be16a2/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:36:55 crc kubenswrapper[4818]: I1203 07:36:55.990893 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-77cd7f4557-lnszf_bcf80b5c-a427-4bcb-a60f-51b730b41cf2/neutron-httpd/0.log" Dec 03 07:36:56 crc kubenswrapper[4818]: I1203 07:36:56.031401 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-77cd7f4557-lnszf_bcf80b5c-a427-4bcb-a60f-51b730b41cf2/neutron-api/0.log" Dec 03 07:36:56 crc kubenswrapper[4818]: I1203 07:36:56.035318 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-487h9_31a69e56-fab5-4dd2-83bb-78cce024c2c1/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:36:56 crc kubenswrapper[4818]: I1203 07:36:56.654009 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_5a5a4f96-9611-456a-a084-88ba9f3b548a/nova-api-log/0.log" Dec 03 07:36:56 crc kubenswrapper[4818]: I1203 07:36:56.735581 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_f40ddc54-764b-4718-aea1-ac8e81eccae7/nova-cell0-conductor-conductor/0.log" Dec 03 07:36:56 crc kubenswrapper[4818]: I1203 07:36:56.943756 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_d772d6d8-be96-4256-84a5-e9f4713ad4ec/nova-cell1-conductor-conductor/0.log" Dec 03 07:36:57 crc kubenswrapper[4818]: I1203 07:36:57.104674 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_b1f0f202-1c95-48ed-96fa-1606b2c81a6f/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 07:36:57 crc kubenswrapper[4818]: I1203 07:36:57.121735 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_5a5a4f96-9611-456a-a084-88ba9f3b548a/nova-api-api/0.log" Dec 03 07:36:57 crc kubenswrapper[4818]: I1203 07:36:57.232495 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-l4l2p_febe4f5e-461e-4562-a168-b99a22ef6e7c/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:36:57 crc kubenswrapper[4818]: I1203 07:36:57.418438 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_78622406-d448-4cf4-b5c9-c639b0fb6a41/nova-metadata-log/0.log" Dec 03 07:36:57 crc kubenswrapper[4818]: I1203 07:36:57.731785 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a0dd3e13-e5ee-4315-ab7a-222c902192be/mysql-bootstrap/0.log" Dec 03 07:36:57 crc kubenswrapper[4818]: I1203 07:36:57.856181 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_39a9a10c-e02f-4468-95a9-651b694b1f9a/nova-scheduler-scheduler/0.log" Dec 03 07:36:57 crc kubenswrapper[4818]: I1203 07:36:57.886930 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a0dd3e13-e5ee-4315-ab7a-222c902192be/galera/0.log" Dec 03 07:36:57 crc kubenswrapper[4818]: I1203 07:36:57.908229 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a0dd3e13-e5ee-4315-ab7a-222c902192be/mysql-bootstrap/0.log" Dec 03 07:36:58 crc kubenswrapper[4818]: I1203 07:36:58.118036 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_206b044a-6264-4c4e-89b9-7cacc4f30ac5/mysql-bootstrap/0.log" Dec 03 07:36:58 crc kubenswrapper[4818]: I1203 07:36:58.316191 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_206b044a-6264-4c4e-89b9-7cacc4f30ac5/galera/0.log" Dec 03 07:36:58 crc kubenswrapper[4818]: I1203 07:36:58.369157 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_206b044a-6264-4c4e-89b9-7cacc4f30ac5/mysql-bootstrap/0.log" Dec 03 07:36:58 crc kubenswrapper[4818]: I1203 07:36:58.537847 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_aecf49dc-5a72-41aa-bc7c-c42bf226ea71/openstackclient/0.log" Dec 03 07:36:58 crc kubenswrapper[4818]: I1203 07:36:58.538792 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-kvg5t_04bf523a-de6a-49a0-93ad-cca1e22939c6/ovn-controller/0.log" Dec 03 07:36:58 crc kubenswrapper[4818]: I1203 07:36:58.741157 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-srrh4_997e3f50-bc25-4436-b03d-5e98770df25f/openstack-network-exporter/0.log" Dec 03 07:36:58 crc kubenswrapper[4818]: I1203 07:36:58.944780 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-brxh4_b5f31813-ba17-4452-9b1a-ee7dcfac564e/ovsdb-server-init/0.log" Dec 03 07:36:59 crc kubenswrapper[4818]: I1203 07:36:59.015849 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_78622406-d448-4cf4-b5c9-c639b0fb6a41/nova-metadata-metadata/0.log" Dec 03 07:36:59 crc kubenswrapper[4818]: I1203 07:36:59.194897 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-brxh4_b5f31813-ba17-4452-9b1a-ee7dcfac564e/ovsdb-server-init/0.log" Dec 03 07:36:59 crc kubenswrapper[4818]: I1203 07:36:59.242546 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-brxh4_b5f31813-ba17-4452-9b1a-ee7dcfac564e/ovs-vswitchd/0.log" Dec 03 07:36:59 crc kubenswrapper[4818]: I1203 07:36:59.247673 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-brxh4_b5f31813-ba17-4452-9b1a-ee7dcfac564e/ovsdb-server/0.log" Dec 03 07:36:59 crc kubenswrapper[4818]: I1203 07:36:59.469372 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-xxxp7_22cbc8e9-1434-4dcb-8fd3-24def21fe451/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:36:59 crc kubenswrapper[4818]: I1203 07:36:59.500256 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8df18ace-e467-4e73-9304-3ea16df510e8/openstack-network-exporter/0.log" Dec 03 07:36:59 crc kubenswrapper[4818]: I1203 07:36:59.542396 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8df18ace-e467-4e73-9304-3ea16df510e8/ovn-northd/0.log" Dec 03 07:37:00 crc kubenswrapper[4818]: I1203 07:37:00.294504 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9b27dc29-202f-4efb-9545-031eaa1d7da3/openstack-network-exporter/0.log" Dec 03 07:37:00 crc kubenswrapper[4818]: I1203 07:37:00.394264 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9b27dc29-202f-4efb-9545-031eaa1d7da3/ovsdbserver-nb/0.log" Dec 03 07:37:00 crc kubenswrapper[4818]: I1203 07:37:00.501136 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_a27ef44f-2bff-4714-8018-a50f2f2de24a/openstack-network-exporter/0.log" Dec 03 07:37:00 crc kubenswrapper[4818]: I1203 07:37:00.627975 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_a27ef44f-2bff-4714-8018-a50f2f2de24a/ovsdbserver-sb/0.log" Dec 03 07:37:00 crc kubenswrapper[4818]: I1203 07:37:00.717033 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-748bc8fb48-scr66_e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff/placement-api/0.log" Dec 03 07:37:00 crc kubenswrapper[4818]: I1203 07:37:00.818842 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-748bc8fb48-scr66_e8fdfb1c-8a82-41cb-a8e0-58e83d5bc8ff/placement-log/0.log" Dec 03 07:37:00 crc kubenswrapper[4818]: I1203 07:37:00.925344 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8078039a-8674-40c1-ba1a-ff8244d17e16/setup-container/0.log" Dec 03 07:37:01 crc kubenswrapper[4818]: I1203 07:37:01.129560 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8078039a-8674-40c1-ba1a-ff8244d17e16/rabbitmq/0.log" Dec 03 07:37:01 crc kubenswrapper[4818]: I1203 07:37:01.133027 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8078039a-8674-40c1-ba1a-ff8244d17e16/setup-container/0.log" Dec 03 07:37:01 crc kubenswrapper[4818]: I1203 07:37:01.147692 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e6fa481e-32ce-4840-b457-0a692558cdb5/setup-container/0.log" Dec 03 07:37:01 crc kubenswrapper[4818]: I1203 07:37:01.460631 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e6fa481e-32ce-4840-b457-0a692558cdb5/rabbitmq/0.log" Dec 03 07:37:01 crc kubenswrapper[4818]: I1203 07:37:01.469244 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e6fa481e-32ce-4840-b457-0a692558cdb5/setup-container/0.log" Dec 03 07:37:01 crc kubenswrapper[4818]: I1203 07:37:01.496944 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-kqb26_3bf68a0c-4bbc-4f9e-b3be-503d857ead66/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:37:01 crc kubenswrapper[4818]: I1203 07:37:01.701605 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-nsd6q_fa6a1813-a4e8-4c7e-bf3b-e3e7d912ca1d/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:37:01 crc kubenswrapper[4818]: I1203 07:37:01.716351 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-b29sh_37db016a-a063-4751-bf39-aadd6746fa13/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:37:02 crc kubenswrapper[4818]: I1203 07:37:02.299570 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-d689l_90add555-879b-4da5-a350-40ce938e9956/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:37:02 crc kubenswrapper[4818]: I1203 07:37:02.350201 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-lbkfx_b054c1a4-49a8-46eb-9e6d-aaaab7cdf2de/ssh-known-hosts-edpm-deployment/0.log" Dec 03 07:37:02 crc kubenswrapper[4818]: I1203 07:37:02.586617 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7954d7ccff-th9lw_0ba50d38-959d-49f4-a98a-6280a78661e4/proxy-server/0.log" Dec 03 07:37:02 crc kubenswrapper[4818]: I1203 07:37:02.715522 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7954d7ccff-th9lw_0ba50d38-959d-49f4-a98a-6280a78661e4/proxy-httpd/0.log" Dec 03 07:37:02 crc kubenswrapper[4818]: I1203 07:37:02.818256 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/account-auditor/0.log" Dec 03 07:37:02 crc kubenswrapper[4818]: I1203 07:37:02.819325 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-pkxxx_337c7d6b-71fc-4044-98ec-de8bfbf5dd6f/swift-ring-rebalance/0.log" Dec 03 07:37:02 crc kubenswrapper[4818]: I1203 07:37:02.965318 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/account-reaper/0.log" Dec 03 07:37:03 crc kubenswrapper[4818]: I1203 07:37:03.061405 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/container-auditor/0.log" Dec 03 07:37:03 crc kubenswrapper[4818]: I1203 07:37:03.108882 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/account-server/0.log" Dec 03 07:37:03 crc kubenswrapper[4818]: I1203 07:37:03.130098 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/account-replicator/0.log" Dec 03 07:37:03 crc kubenswrapper[4818]: I1203 07:37:03.196004 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/container-replicator/0.log" Dec 03 07:37:03 crc kubenswrapper[4818]: I1203 07:37:03.268366 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/container-server/0.log" Dec 03 07:37:03 crc kubenswrapper[4818]: I1203 07:37:03.335372 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/container-updater/0.log" Dec 03 07:37:03 crc kubenswrapper[4818]: I1203 07:37:03.350714 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/object-auditor/0.log" Dec 03 07:37:03 crc kubenswrapper[4818]: I1203 07:37:03.392214 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/object-expirer/0.log" Dec 03 07:37:03 crc kubenswrapper[4818]: I1203 07:37:03.643384 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/object-replicator/0.log" Dec 03 07:37:03 crc kubenswrapper[4818]: I1203 07:37:03.661868 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/object-server/0.log" Dec 03 07:37:03 crc kubenswrapper[4818]: I1203 07:37:03.743373 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/object-updater/0.log" Dec 03 07:37:03 crc kubenswrapper[4818]: I1203 07:37:03.753401 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/rsync/0.log" Dec 03 07:37:03 crc kubenswrapper[4818]: I1203 07:37:03.887059 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_520a6b29-d2d4-46e4-a5ed-989f2fa1df1f/swift-recon-cron/0.log" Dec 03 07:37:03 crc kubenswrapper[4818]: I1203 07:37:03.995731 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-wkr94_9113c780-f00c-4f24-b000-26270c809028/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:37:04 crc kubenswrapper[4818]: I1203 07:37:04.136647 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_a43ff7a5-52b3-4e12-a9c6-ec3b5b8bd8cf/tempest-tests-tempest-tests-runner/0.log" Dec 03 07:37:04 crc kubenswrapper[4818]: I1203 07:37:04.496947 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_78a39ae7-b979-4d32-a128-25dca0a393e4/test-operator-logs-container/0.log" Dec 03 07:37:04 crc kubenswrapper[4818]: I1203 07:37:04.521288 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-tkmbh_4e2bf3bf-bd96-4042-b1e6-11247aefdbe2/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 07:37:14 crc kubenswrapper[4818]: I1203 07:37:14.135264 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_c826858d-45cb-4ae3-8fa2-1531f9390480/memcached/0.log" Dec 03 07:37:33 crc kubenswrapper[4818]: I1203 07:37:33.177269 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-z6c4f_7c36b14e-4e1d-4956-a505-ecdffb9d9f9c/kube-rbac-proxy/0.log" Dec 03 07:37:33 crc kubenswrapper[4818]: I1203 07:37:33.243569 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-z6c4f_7c36b14e-4e1d-4956-a505-ecdffb9d9f9c/manager/0.log" Dec 03 07:37:33 crc kubenswrapper[4818]: I1203 07:37:33.366480 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-4hdr6_4130e7e0-4596-451f-b346-5afba08e9dea/kube-rbac-proxy/0.log" Dec 03 07:37:33 crc kubenswrapper[4818]: I1203 07:37:33.418009 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-4hdr6_4130e7e0-4596-451f-b346-5afba08e9dea/manager/0.log" Dec 03 07:37:33 crc kubenswrapper[4818]: I1203 07:37:33.529861 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk_413ec55a-90aa-4e14-a447-47595f913b9a/util/0.log" Dec 03 07:37:33 crc kubenswrapper[4818]: I1203 07:37:33.696806 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk_413ec55a-90aa-4e14-a447-47595f913b9a/pull/0.log" Dec 03 07:37:33 crc kubenswrapper[4818]: I1203 07:37:33.719992 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk_413ec55a-90aa-4e14-a447-47595f913b9a/util/0.log" Dec 03 07:37:33 crc kubenswrapper[4818]: I1203 07:37:33.720541 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk_413ec55a-90aa-4e14-a447-47595f913b9a/pull/0.log" Dec 03 07:37:33 crc kubenswrapper[4818]: I1203 07:37:33.884580 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk_413ec55a-90aa-4e14-a447-47595f913b9a/pull/0.log" Dec 03 07:37:33 crc kubenswrapper[4818]: I1203 07:37:33.885026 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk_413ec55a-90aa-4e14-a447-47595f913b9a/util/0.log" Dec 03 07:37:33 crc kubenswrapper[4818]: I1203 07:37:33.923378 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d2b89cb9716f44b9c475cfb7e71244f393fe86fe2707a7d4113f58c65agkdvk_413ec55a-90aa-4e14-a447-47595f913b9a/extract/0.log" Dec 03 07:37:34 crc kubenswrapper[4818]: I1203 07:37:34.057685 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-wndt9_91e645ae-28ec-4d3b-8fe3-e75aa1acf5a5/kube-rbac-proxy/0.log" Dec 03 07:37:34 crc kubenswrapper[4818]: I1203 07:37:34.093617 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-wndt9_91e645ae-28ec-4d3b-8fe3-e75aa1acf5a5/manager/0.log" Dec 03 07:37:34 crc kubenswrapper[4818]: I1203 07:37:34.128172 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-2qvsd_528c7c89-7980-4acc-9321-ff56c9d38489/kube-rbac-proxy/0.log" Dec 03 07:37:34 crc kubenswrapper[4818]: I1203 07:37:34.318143 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-gtmmz_b29acffb-a7e4-4029-93fa-20f20de27dfe/kube-rbac-proxy/0.log" Dec 03 07:37:34 crc kubenswrapper[4818]: I1203 07:37:34.325778 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-2qvsd_528c7c89-7980-4acc-9321-ff56c9d38489/manager/0.log" Dec 03 07:37:34 crc kubenswrapper[4818]: I1203 07:37:34.375586 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-gtmmz_b29acffb-a7e4-4029-93fa-20f20de27dfe/manager/0.log" Dec 03 07:37:34 crc kubenswrapper[4818]: I1203 07:37:34.504140 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-thzjt_6d5bf0d5-df89-4ab4-a711-e6f7f3d1123e/kube-rbac-proxy/0.log" Dec 03 07:37:34 crc kubenswrapper[4818]: I1203 07:37:34.547498 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-thzjt_6d5bf0d5-df89-4ab4-a711-e6f7f3d1123e/manager/0.log" Dec 03 07:37:34 crc kubenswrapper[4818]: I1203 07:37:34.686141 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-gmkzd_a0b8fddb-4636-4f13-8d0b-3e3220721eca/kube-rbac-proxy/0.log" Dec 03 07:37:34 crc kubenswrapper[4818]: I1203 07:37:34.802267 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-hpnhz_3d1565f0-62d4-4ecd-9d73-16338035196e/kube-rbac-proxy/0.log" Dec 03 07:37:34 crc kubenswrapper[4818]: I1203 07:37:34.840376 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-gmkzd_a0b8fddb-4636-4f13-8d0b-3e3220721eca/manager/0.log" Dec 03 07:37:34 crc kubenswrapper[4818]: I1203 07:37:34.886764 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-hpnhz_3d1565f0-62d4-4ecd-9d73-16338035196e/manager/0.log" Dec 03 07:37:34 crc kubenswrapper[4818]: I1203 07:37:34.995915 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-27nrr_1bd3fb88-7598-4a28-90fe-324d855dc90e/kube-rbac-proxy/0.log" Dec 03 07:37:35 crc kubenswrapper[4818]: I1203 07:37:35.058408 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-27nrr_1bd3fb88-7598-4a28-90fe-324d855dc90e/manager/0.log" Dec 03 07:37:35 crc kubenswrapper[4818]: I1203 07:37:35.151933 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-s84mt_0c360185-7fff-4054-a85e-ca2bbb1fa5e0/kube-rbac-proxy/0.log" Dec 03 07:37:35 crc kubenswrapper[4818]: I1203 07:37:35.186147 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-s84mt_0c360185-7fff-4054-a85e-ca2bbb1fa5e0/manager/0.log" Dec 03 07:37:35 crc kubenswrapper[4818]: I1203 07:37:35.316088 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-98z98_df292cb6-9f13-4cec-a7e1-1ef3155a7b1f/kube-rbac-proxy/0.log" Dec 03 07:37:35 crc kubenswrapper[4818]: I1203 07:37:35.362250 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-98z98_df292cb6-9f13-4cec-a7e1-1ef3155a7b1f/manager/0.log" Dec 03 07:37:35 crc kubenswrapper[4818]: I1203 07:37:35.481401 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-b4thx_0d4b3be5-bb2e-4580-8142-60d15b16c36f/kube-rbac-proxy/0.log" Dec 03 07:37:35 crc kubenswrapper[4818]: I1203 07:37:35.588410 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-b4thx_0d4b3be5-bb2e-4580-8142-60d15b16c36f/manager/0.log" Dec 03 07:37:35 crc kubenswrapper[4818]: I1203 07:37:35.741640 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-xtq4s_1dc6c089-07b6-4818-a138-6b6a8c0cd1b5/kube-rbac-proxy/0.log" Dec 03 07:37:35 crc kubenswrapper[4818]: I1203 07:37:35.755928 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-xtq4s_1dc6c089-07b6-4818-a138-6b6a8c0cd1b5/manager/0.log" Dec 03 07:37:35 crc kubenswrapper[4818]: I1203 07:37:35.791908 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-6rr64_45dcf421-464d-4eb9-a23f-9924ecf5bb03/kube-rbac-proxy/0.log" Dec 03 07:37:36 crc kubenswrapper[4818]: I1203 07:37:36.077337 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-6rr64_45dcf421-464d-4eb9-a23f-9924ecf5bb03/manager/0.log" Dec 03 07:37:36 crc kubenswrapper[4818]: I1203 07:37:36.139392 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd445pgf_f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34/kube-rbac-proxy/0.log" Dec 03 07:37:36 crc kubenswrapper[4818]: I1203 07:37:36.152288 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd445pgf_f24eeb0d-5a36-4dd7-93e8-3e87f1fa7e34/manager/0.log" Dec 03 07:37:36 crc kubenswrapper[4818]: I1203 07:37:36.554214 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-jr8dk_f28e0eba-98e8-440c-8c04-91ce47ea304b/registry-server/0.log" Dec 03 07:37:36 crc kubenswrapper[4818]: I1203 07:37:36.590501 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-555856bbb4-tlxbc_69cb7eae-cb67-42ef-b36d-d6287f1f14d4/operator/0.log" Dec 03 07:37:36 crc kubenswrapper[4818]: I1203 07:37:36.822449 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-nchr7_4b33f4e2-edb8-4e90-b73a-4dc2b13edd41/kube-rbac-proxy/0.log" Dec 03 07:37:36 crc kubenswrapper[4818]: I1203 07:37:36.874435 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-nchr7_4b33f4e2-edb8-4e90-b73a-4dc2b13edd41/manager/0.log" Dec 03 07:37:36 crc kubenswrapper[4818]: I1203 07:37:36.992612 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-2d5xh_e977170c-c8e9-4888-951b-bffa4d4d21d7/kube-rbac-proxy/0.log" Dec 03 07:37:37 crc kubenswrapper[4818]: I1203 07:37:37.114053 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-2d5xh_e977170c-c8e9-4888-951b-bffa4d4d21d7/manager/0.log" Dec 03 07:37:37 crc kubenswrapper[4818]: I1203 07:37:37.302259 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-r4b9n_2e8a3376-fa5d-458d-a659-0c3cef8af10b/kube-rbac-proxy/0.log" Dec 03 07:37:37 crc kubenswrapper[4818]: I1203 07:37:37.351236 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-5q95v_5a1859a2-fa2e-4d8c-bd64-7b63297f258c/operator/0.log" Dec 03 07:37:37 crc kubenswrapper[4818]: I1203 07:37:37.411018 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-r4b9n_2e8a3376-fa5d-458d-a659-0c3cef8af10b/manager/0.log" Dec 03 07:37:37 crc kubenswrapper[4818]: I1203 07:37:37.525955 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-98d87bf44-8h8d6_3a635982-10c6-47cc-8462-74873b98410a/manager/0.log" Dec 03 07:37:37 crc kubenswrapper[4818]: I1203 07:37:37.599260 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-bjq7j_b78bb8f0-22f9-4b8e-93e2-23422db45354/kube-rbac-proxy/0.log" Dec 03 07:37:37 crc kubenswrapper[4818]: I1203 07:37:37.623916 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-bjq7j_b78bb8f0-22f9-4b8e-93e2-23422db45354/manager/0.log" Dec 03 07:37:37 crc kubenswrapper[4818]: I1203 07:37:37.723120 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-8hntr_84dd0135-3ec6-4418-98f9-de8474662893/kube-rbac-proxy/0.log" Dec 03 07:37:37 crc kubenswrapper[4818]: I1203 07:37:37.754180 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-8hntr_84dd0135-3ec6-4418-98f9-de8474662893/manager/0.log" Dec 03 07:37:37 crc kubenswrapper[4818]: I1203 07:37:37.808781 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-jd5hg_56bb5f70-9dd8-433f-a974-0c95065c1dd1/kube-rbac-proxy/0.log" Dec 03 07:37:37 crc kubenswrapper[4818]: I1203 07:37:37.867333 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-jd5hg_56bb5f70-9dd8-433f-a974-0c95065c1dd1/manager/0.log" Dec 03 07:37:56 crc kubenswrapper[4818]: I1203 07:37:56.567842 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-ltb69_faad9125-f494-49ae-ba77-941e6abe67b9/control-plane-machine-set-operator/0.log" Dec 03 07:37:56 crc kubenswrapper[4818]: I1203 07:37:56.684402 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jf227_8ec6e899-8380-4812-8f65-6fd72db12939/kube-rbac-proxy/0.log" Dec 03 07:37:56 crc kubenswrapper[4818]: I1203 07:37:56.765455 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jf227_8ec6e899-8380-4812-8f65-6fd72db12939/machine-api-operator/0.log" Dec 03 07:38:05 crc kubenswrapper[4818]: I1203 07:38:05.726741 4818 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4wkmv"] Dec 03 07:38:05 crc kubenswrapper[4818]: E1203 07:38:05.729037 4818 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4acb60dd-6f98-4c73-b65b-31393936d2f9" containerName="container-00" Dec 03 07:38:05 crc kubenswrapper[4818]: I1203 07:38:05.729057 4818 state_mem.go:107] "Deleted CPUSet assignment" podUID="4acb60dd-6f98-4c73-b65b-31393936d2f9" containerName="container-00" Dec 03 07:38:05 crc kubenswrapper[4818]: I1203 07:38:05.729374 4818 memory_manager.go:354] "RemoveStaleState removing state" podUID="4acb60dd-6f98-4c73-b65b-31393936d2f9" containerName="container-00" Dec 03 07:38:05 crc kubenswrapper[4818]: I1203 07:38:05.731110 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4wkmv" Dec 03 07:38:05 crc kubenswrapper[4818]: I1203 07:38:05.739733 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4wkmv"] Dec 03 07:38:05 crc kubenswrapper[4818]: I1203 07:38:05.786923 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bba00df-133e-484c-a16e-b2769ca31b7e-catalog-content\") pod \"community-operators-4wkmv\" (UID: \"6bba00df-133e-484c-a16e-b2769ca31b7e\") " pod="openshift-marketplace/community-operators-4wkmv" Dec 03 07:38:05 crc kubenswrapper[4818]: I1203 07:38:05.787043 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw4rm\" (UniqueName: \"kubernetes.io/projected/6bba00df-133e-484c-a16e-b2769ca31b7e-kube-api-access-pw4rm\") pod \"community-operators-4wkmv\" (UID: \"6bba00df-133e-484c-a16e-b2769ca31b7e\") " pod="openshift-marketplace/community-operators-4wkmv" Dec 03 07:38:05 crc kubenswrapper[4818]: I1203 07:38:05.787558 4818 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bba00df-133e-484c-a16e-b2769ca31b7e-utilities\") pod \"community-operators-4wkmv\" (UID: \"6bba00df-133e-484c-a16e-b2769ca31b7e\") " pod="openshift-marketplace/community-operators-4wkmv" Dec 03 07:38:05 crc kubenswrapper[4818]: I1203 07:38:05.889942 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw4rm\" (UniqueName: \"kubernetes.io/projected/6bba00df-133e-484c-a16e-b2769ca31b7e-kube-api-access-pw4rm\") pod \"community-operators-4wkmv\" (UID: \"6bba00df-133e-484c-a16e-b2769ca31b7e\") " pod="openshift-marketplace/community-operators-4wkmv" Dec 03 07:38:05 crc kubenswrapper[4818]: I1203 07:38:05.890339 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bba00df-133e-484c-a16e-b2769ca31b7e-utilities\") pod \"community-operators-4wkmv\" (UID: \"6bba00df-133e-484c-a16e-b2769ca31b7e\") " pod="openshift-marketplace/community-operators-4wkmv" Dec 03 07:38:05 crc kubenswrapper[4818]: I1203 07:38:05.890386 4818 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bba00df-133e-484c-a16e-b2769ca31b7e-catalog-content\") pod \"community-operators-4wkmv\" (UID: \"6bba00df-133e-484c-a16e-b2769ca31b7e\") " pod="openshift-marketplace/community-operators-4wkmv" Dec 03 07:38:05 crc kubenswrapper[4818]: I1203 07:38:05.890973 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bba00df-133e-484c-a16e-b2769ca31b7e-catalog-content\") pod \"community-operators-4wkmv\" (UID: \"6bba00df-133e-484c-a16e-b2769ca31b7e\") " pod="openshift-marketplace/community-operators-4wkmv" Dec 03 07:38:05 crc kubenswrapper[4818]: I1203 07:38:05.890970 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bba00df-133e-484c-a16e-b2769ca31b7e-utilities\") pod \"community-operators-4wkmv\" (UID: \"6bba00df-133e-484c-a16e-b2769ca31b7e\") " pod="openshift-marketplace/community-operators-4wkmv" Dec 03 07:38:05 crc kubenswrapper[4818]: I1203 07:38:05.913721 4818 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw4rm\" (UniqueName: \"kubernetes.io/projected/6bba00df-133e-484c-a16e-b2769ca31b7e-kube-api-access-pw4rm\") pod \"community-operators-4wkmv\" (UID: \"6bba00df-133e-484c-a16e-b2769ca31b7e\") " pod="openshift-marketplace/community-operators-4wkmv" Dec 03 07:38:06 crc kubenswrapper[4818]: I1203 07:38:06.052940 4818 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4wkmv" Dec 03 07:38:06 crc kubenswrapper[4818]: I1203 07:38:06.578442 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4wkmv"] Dec 03 07:38:07 crc kubenswrapper[4818]: I1203 07:38:07.006134 4818 generic.go:334] "Generic (PLEG): container finished" podID="6bba00df-133e-484c-a16e-b2769ca31b7e" containerID="64e280711d3873fdbb25471d11e661ae8b84e6362c8bd9af6f09a0f9ebcb65cb" exitCode=0 Dec 03 07:38:07 crc kubenswrapper[4818]: I1203 07:38:07.006256 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4wkmv" event={"ID":"6bba00df-133e-484c-a16e-b2769ca31b7e","Type":"ContainerDied","Data":"64e280711d3873fdbb25471d11e661ae8b84e6362c8bd9af6f09a0f9ebcb65cb"} Dec 03 07:38:07 crc kubenswrapper[4818]: I1203 07:38:07.006697 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4wkmv" event={"ID":"6bba00df-133e-484c-a16e-b2769ca31b7e","Type":"ContainerStarted","Data":"0ea0d2c8d2851b7b474c9fafa609627110bc158f111d68deaff96033f4e35214"} Dec 03 07:38:08 crc kubenswrapper[4818]: I1203 07:38:08.823073 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-xl5nj_47f0d000-e043-4e9e-b300-988f03341b42/cert-manager-controller/0.log" Dec 03 07:38:09 crc kubenswrapper[4818]: I1203 07:38:09.076635 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-tcn75_8f6981d3-9529-498d-935e-f99f533faa9e/cert-manager-webhook/0.log" Dec 03 07:38:09 crc kubenswrapper[4818]: I1203 07:38:09.097952 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-cpcd5_c9bd0ac7-83ba-4f70-9458-0d8b88618456/cert-manager-cainjector/0.log" Dec 03 07:38:12 crc kubenswrapper[4818]: I1203 07:38:12.061103 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4wkmv" event={"ID":"6bba00df-133e-484c-a16e-b2769ca31b7e","Type":"ContainerStarted","Data":"6d617c16ef02218c640b7682c9ac0322c38f5c19186552ec035db46655ca3bbb"} Dec 03 07:38:13 crc kubenswrapper[4818]: I1203 07:38:13.302023 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:38:13 crc kubenswrapper[4818]: I1203 07:38:13.302368 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:38:14 crc kubenswrapper[4818]: I1203 07:38:14.087709 4818 generic.go:334] "Generic (PLEG): container finished" podID="6bba00df-133e-484c-a16e-b2769ca31b7e" containerID="6d617c16ef02218c640b7682c9ac0322c38f5c19186552ec035db46655ca3bbb" exitCode=0 Dec 03 07:38:14 crc kubenswrapper[4818]: I1203 07:38:14.087762 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4wkmv" event={"ID":"6bba00df-133e-484c-a16e-b2769ca31b7e","Type":"ContainerDied","Data":"6d617c16ef02218c640b7682c9ac0322c38f5c19186552ec035db46655ca3bbb"} Dec 03 07:38:15 crc kubenswrapper[4818]: I1203 07:38:15.100144 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4wkmv" event={"ID":"6bba00df-133e-484c-a16e-b2769ca31b7e","Type":"ContainerStarted","Data":"4dcfca7fe4582a58d1eaed1d569baf19e28cfebc55e028bf7083ed0048e1dce3"} Dec 03 07:38:15 crc kubenswrapper[4818]: I1203 07:38:15.131571 4818 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4wkmv" podStartSLOduration=2.520341594 podStartE2EDuration="10.131552043s" podCreationTimestamp="2025-12-03 07:38:05 +0000 UTC" firstStartedPulling="2025-12-03 07:38:07.007690079 +0000 UTC m=+4244.699298831" lastFinishedPulling="2025-12-03 07:38:14.618900538 +0000 UTC m=+4252.310509280" observedRunningTime="2025-12-03 07:38:15.124761084 +0000 UTC m=+4252.816369846" watchObservedRunningTime="2025-12-03 07:38:15.131552043 +0000 UTC m=+4252.823160805" Dec 03 07:38:16 crc kubenswrapper[4818]: I1203 07:38:16.053166 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4wkmv" Dec 03 07:38:16 crc kubenswrapper[4818]: I1203 07:38:16.053231 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4wkmv" Dec 03 07:38:16 crc kubenswrapper[4818]: I1203 07:38:16.253508 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-7954d7ccff-th9lw" podUID="0ba50d38-959d-49f4-a98a-6280a78661e4" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 03 07:38:17 crc kubenswrapper[4818]: I1203 07:38:17.403069 4818 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-4wkmv" podUID="6bba00df-133e-484c-a16e-b2769ca31b7e" containerName="registry-server" probeResult="failure" output=< Dec 03 07:38:17 crc kubenswrapper[4818]: timeout: failed to connect service ":50051" within 1s Dec 03 07:38:17 crc kubenswrapper[4818]: > Dec 03 07:38:22 crc kubenswrapper[4818]: I1203 07:38:22.806922 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-ncmjw_3c885e5f-3396-4927-90ef-878b255a6317/nmstate-console-plugin/0.log" Dec 03 07:38:22 crc kubenswrapper[4818]: I1203 07:38:22.964342 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-mrrpc_2ec105b0-3a09-4c8d-a494-95116b4329f2/kube-rbac-proxy/0.log" Dec 03 07:38:23 crc kubenswrapper[4818]: I1203 07:38:23.007289 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-mrrpc_2ec105b0-3a09-4c8d-a494-95116b4329f2/nmstate-metrics/0.log" Dec 03 07:38:23 crc kubenswrapper[4818]: I1203 07:38:23.014440 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-7hwmb_1d3707ea-c681-45e2-959f-336f11ad216d/nmstate-handler/0.log" Dec 03 07:38:23 crc kubenswrapper[4818]: I1203 07:38:23.212897 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-phlrm_160775bb-417d-4852-bc1f-77d6f8542fad/nmstate-operator/0.log" Dec 03 07:38:24 crc kubenswrapper[4818]: I1203 07:38:24.150688 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-lp854_fbefe4c2-3aea-4974-941e-74fc9448cf32/nmstate-webhook/0.log" Dec 03 07:38:26 crc kubenswrapper[4818]: I1203 07:38:26.099022 4818 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4wkmv" Dec 03 07:38:26 crc kubenswrapper[4818]: I1203 07:38:26.148394 4818 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4wkmv" Dec 03 07:38:26 crc kubenswrapper[4818]: I1203 07:38:26.266516 4818 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4wkmv"] Dec 03 07:38:26 crc kubenswrapper[4818]: I1203 07:38:26.365377 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mwgrm"] Dec 03 07:38:26 crc kubenswrapper[4818]: I1203 07:38:26.365707 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mwgrm" podUID="57e20fce-c349-473f-8866-b194da09c27d" containerName="registry-server" containerID="cri-o://afde14fa7c5a8f50dd12f8f9c872f17a9f54cfda6fe8090cd71eb8fb5744ec4a" gracePeriod=2 Dec 03 07:38:26 crc kubenswrapper[4818]: I1203 07:38:26.851949 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mwgrm" Dec 03 07:38:26 crc kubenswrapper[4818]: I1203 07:38:26.963137 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zq4z4\" (UniqueName: \"kubernetes.io/projected/57e20fce-c349-473f-8866-b194da09c27d-kube-api-access-zq4z4\") pod \"57e20fce-c349-473f-8866-b194da09c27d\" (UID: \"57e20fce-c349-473f-8866-b194da09c27d\") " Dec 03 07:38:26 crc kubenswrapper[4818]: I1203 07:38:26.963199 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57e20fce-c349-473f-8866-b194da09c27d-catalog-content\") pod \"57e20fce-c349-473f-8866-b194da09c27d\" (UID: \"57e20fce-c349-473f-8866-b194da09c27d\") " Dec 03 07:38:26 crc kubenswrapper[4818]: I1203 07:38:26.963303 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57e20fce-c349-473f-8866-b194da09c27d-utilities\") pod \"57e20fce-c349-473f-8866-b194da09c27d\" (UID: \"57e20fce-c349-473f-8866-b194da09c27d\") " Dec 03 07:38:26 crc kubenswrapper[4818]: I1203 07:38:26.965929 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57e20fce-c349-473f-8866-b194da09c27d-utilities" (OuterVolumeSpecName: "utilities") pod "57e20fce-c349-473f-8866-b194da09c27d" (UID: "57e20fce-c349-473f-8866-b194da09c27d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:38:26 crc kubenswrapper[4818]: I1203 07:38:26.975402 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57e20fce-c349-473f-8866-b194da09c27d-kube-api-access-zq4z4" (OuterVolumeSpecName: "kube-api-access-zq4z4") pod "57e20fce-c349-473f-8866-b194da09c27d" (UID: "57e20fce-c349-473f-8866-b194da09c27d"). InnerVolumeSpecName "kube-api-access-zq4z4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.048410 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57e20fce-c349-473f-8866-b194da09c27d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57e20fce-c349-473f-8866-b194da09c27d" (UID: "57e20fce-c349-473f-8866-b194da09c27d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.067563 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zq4z4\" (UniqueName: \"kubernetes.io/projected/57e20fce-c349-473f-8866-b194da09c27d-kube-api-access-zq4z4\") on node \"crc\" DevicePath \"\"" Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.067597 4818 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57e20fce-c349-473f-8866-b194da09c27d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.067606 4818 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57e20fce-c349-473f-8866-b194da09c27d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.222393 4818 generic.go:334] "Generic (PLEG): container finished" podID="57e20fce-c349-473f-8866-b194da09c27d" containerID="afde14fa7c5a8f50dd12f8f9c872f17a9f54cfda6fe8090cd71eb8fb5744ec4a" exitCode=0 Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.222447 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwgrm" event={"ID":"57e20fce-c349-473f-8866-b194da09c27d","Type":"ContainerDied","Data":"afde14fa7c5a8f50dd12f8f9c872f17a9f54cfda6fe8090cd71eb8fb5744ec4a"} Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.222488 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mwgrm" Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.222525 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mwgrm" event={"ID":"57e20fce-c349-473f-8866-b194da09c27d","Type":"ContainerDied","Data":"f4b652b6c084a48ff4d790896af57ad3bc2ffdef97bddd472376140d61b98952"} Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.222593 4818 scope.go:117] "RemoveContainer" containerID="afde14fa7c5a8f50dd12f8f9c872f17a9f54cfda6fe8090cd71eb8fb5744ec4a" Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.246145 4818 scope.go:117] "RemoveContainer" containerID="a997770edc0889bbee6d1b674b13edd67f1631820f46dd673b3e53e9359a087b" Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.281147 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mwgrm"] Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.288013 4818 scope.go:117] "RemoveContainer" containerID="2aae74fe42431c099446f21e6a21ed6734d3888258544577eb1b4997855e210f" Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.294855 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mwgrm"] Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.331570 4818 scope.go:117] "RemoveContainer" containerID="afde14fa7c5a8f50dd12f8f9c872f17a9f54cfda6fe8090cd71eb8fb5744ec4a" Dec 03 07:38:27 crc kubenswrapper[4818]: E1203 07:38:27.332210 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afde14fa7c5a8f50dd12f8f9c872f17a9f54cfda6fe8090cd71eb8fb5744ec4a\": container with ID starting with afde14fa7c5a8f50dd12f8f9c872f17a9f54cfda6fe8090cd71eb8fb5744ec4a not found: ID does not exist" containerID="afde14fa7c5a8f50dd12f8f9c872f17a9f54cfda6fe8090cd71eb8fb5744ec4a" Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.332264 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afde14fa7c5a8f50dd12f8f9c872f17a9f54cfda6fe8090cd71eb8fb5744ec4a"} err="failed to get container status \"afde14fa7c5a8f50dd12f8f9c872f17a9f54cfda6fe8090cd71eb8fb5744ec4a\": rpc error: code = NotFound desc = could not find container \"afde14fa7c5a8f50dd12f8f9c872f17a9f54cfda6fe8090cd71eb8fb5744ec4a\": container with ID starting with afde14fa7c5a8f50dd12f8f9c872f17a9f54cfda6fe8090cd71eb8fb5744ec4a not found: ID does not exist" Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.332293 4818 scope.go:117] "RemoveContainer" containerID="a997770edc0889bbee6d1b674b13edd67f1631820f46dd673b3e53e9359a087b" Dec 03 07:38:27 crc kubenswrapper[4818]: E1203 07:38:27.333201 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a997770edc0889bbee6d1b674b13edd67f1631820f46dd673b3e53e9359a087b\": container with ID starting with a997770edc0889bbee6d1b674b13edd67f1631820f46dd673b3e53e9359a087b not found: ID does not exist" containerID="a997770edc0889bbee6d1b674b13edd67f1631820f46dd673b3e53e9359a087b" Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.333243 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a997770edc0889bbee6d1b674b13edd67f1631820f46dd673b3e53e9359a087b"} err="failed to get container status \"a997770edc0889bbee6d1b674b13edd67f1631820f46dd673b3e53e9359a087b\": rpc error: code = NotFound desc = could not find container \"a997770edc0889bbee6d1b674b13edd67f1631820f46dd673b3e53e9359a087b\": container with ID starting with a997770edc0889bbee6d1b674b13edd67f1631820f46dd673b3e53e9359a087b not found: ID does not exist" Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.333264 4818 scope.go:117] "RemoveContainer" containerID="2aae74fe42431c099446f21e6a21ed6734d3888258544577eb1b4997855e210f" Dec 03 07:38:27 crc kubenswrapper[4818]: E1203 07:38:27.333516 4818 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2aae74fe42431c099446f21e6a21ed6734d3888258544577eb1b4997855e210f\": container with ID starting with 2aae74fe42431c099446f21e6a21ed6734d3888258544577eb1b4997855e210f not found: ID does not exist" containerID="2aae74fe42431c099446f21e6a21ed6734d3888258544577eb1b4997855e210f" Dec 03 07:38:27 crc kubenswrapper[4818]: I1203 07:38:27.333542 4818 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2aae74fe42431c099446f21e6a21ed6734d3888258544577eb1b4997855e210f"} err="failed to get container status \"2aae74fe42431c099446f21e6a21ed6734d3888258544577eb1b4997855e210f\": rpc error: code = NotFound desc = could not find container \"2aae74fe42431c099446f21e6a21ed6734d3888258544577eb1b4997855e210f\": container with ID starting with 2aae74fe42431c099446f21e6a21ed6734d3888258544577eb1b4997855e210f not found: ID does not exist" Dec 03 07:38:28 crc kubenswrapper[4818]: I1203 07:38:28.748557 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57e20fce-c349-473f-8866-b194da09c27d" path="/var/lib/kubelet/pods/57e20fce-c349-473f-8866-b194da09c27d/volumes" Dec 03 07:38:40 crc kubenswrapper[4818]: I1203 07:38:40.678177 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-7nr97_bf628f10-0970-4651-8499-5302e78c8d6d/kube-rbac-proxy/0.log" Dec 03 07:38:40 crc kubenswrapper[4818]: I1203 07:38:40.760182 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-7nr97_bf628f10-0970-4651-8499-5302e78c8d6d/controller/0.log" Dec 03 07:38:40 crc kubenswrapper[4818]: I1203 07:38:40.863593 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-frr-files/0.log" Dec 03 07:38:41 crc kubenswrapper[4818]: I1203 07:38:41.031194 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-frr-files/0.log" Dec 03 07:38:41 crc kubenswrapper[4818]: I1203 07:38:41.049327 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-reloader/0.log" Dec 03 07:38:41 crc kubenswrapper[4818]: I1203 07:38:41.060344 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-reloader/0.log" Dec 03 07:38:41 crc kubenswrapper[4818]: I1203 07:38:41.098395 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-metrics/0.log" Dec 03 07:38:41 crc kubenswrapper[4818]: I1203 07:38:41.450677 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-reloader/0.log" Dec 03 07:38:41 crc kubenswrapper[4818]: I1203 07:38:41.451187 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-metrics/0.log" Dec 03 07:38:41 crc kubenswrapper[4818]: I1203 07:38:41.458673 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-frr-files/0.log" Dec 03 07:38:41 crc kubenswrapper[4818]: I1203 07:38:41.525530 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-metrics/0.log" Dec 03 07:38:41 crc kubenswrapper[4818]: I1203 07:38:41.665115 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-frr-files/0.log" Dec 03 07:38:41 crc kubenswrapper[4818]: I1203 07:38:41.691614 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-reloader/0.log" Dec 03 07:38:41 crc kubenswrapper[4818]: I1203 07:38:41.707025 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/controller/0.log" Dec 03 07:38:41 crc kubenswrapper[4818]: I1203 07:38:41.707298 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/cp-metrics/0.log" Dec 03 07:38:41 crc kubenswrapper[4818]: I1203 07:38:41.874534 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/frr-metrics/0.log" Dec 03 07:38:41 crc kubenswrapper[4818]: I1203 07:38:41.907181 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/kube-rbac-proxy/0.log" Dec 03 07:38:41 crc kubenswrapper[4818]: I1203 07:38:41.972721 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/kube-rbac-proxy-frr/0.log" Dec 03 07:38:42 crc kubenswrapper[4818]: I1203 07:38:42.100742 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/reloader/0.log" Dec 03 07:38:42 crc kubenswrapper[4818]: I1203 07:38:42.304128 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-tkg7w_f55ffc30-37c1-45e8-8e63-c447ca4d6e96/frr-k8s-webhook-server/0.log" Dec 03 07:38:42 crc kubenswrapper[4818]: I1203 07:38:42.972125 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5545bb6cbd-gjkwk_d227f431-7cf6-43ed-85d0-6e44812df281/manager/0.log" Dec 03 07:38:42 crc kubenswrapper[4818]: I1203 07:38:42.992414 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-56f4c8658f-s8ghc_4563221c-8432-4038-b0e6-fdbac56b99e0/webhook-server/0.log" Dec 03 07:38:43 crc kubenswrapper[4818]: I1203 07:38:43.228153 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hwmtn_0a528d4d-3d84-40c2-a4b3-207001ba3579/frr/0.log" Dec 03 07:38:43 crc kubenswrapper[4818]: I1203 07:38:43.248137 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-4m9n7_860c17a3-a17f-4ac3-9310-c633df611f6d/kube-rbac-proxy/0.log" Dec 03 07:38:43 crc kubenswrapper[4818]: I1203 07:38:43.301783 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:38:43 crc kubenswrapper[4818]: I1203 07:38:43.301854 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:38:43 crc kubenswrapper[4818]: I1203 07:38:43.566883 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-4m9n7_860c17a3-a17f-4ac3-9310-c633df611f6d/speaker/0.log" Dec 03 07:38:56 crc kubenswrapper[4818]: I1203 07:38:56.869251 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5_37089b98-d9bb-4265-9e86-eeb0e4e0be99/util/0.log" Dec 03 07:38:57 crc kubenswrapper[4818]: I1203 07:38:57.020112 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5_37089b98-d9bb-4265-9e86-eeb0e4e0be99/util/0.log" Dec 03 07:38:57 crc kubenswrapper[4818]: I1203 07:38:57.066535 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5_37089b98-d9bb-4265-9e86-eeb0e4e0be99/pull/0.log" Dec 03 07:38:57 crc kubenswrapper[4818]: I1203 07:38:57.123252 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5_37089b98-d9bb-4265-9e86-eeb0e4e0be99/pull/0.log" Dec 03 07:38:57 crc kubenswrapper[4818]: I1203 07:38:57.279511 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5_37089b98-d9bb-4265-9e86-eeb0e4e0be99/pull/0.log" Dec 03 07:38:57 crc kubenswrapper[4818]: I1203 07:38:57.322018 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5_37089b98-d9bb-4265-9e86-eeb0e4e0be99/util/0.log" Dec 03 07:38:57 crc kubenswrapper[4818]: I1203 07:38:57.323010 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9kzq5_37089b98-d9bb-4265-9e86-eeb0e4e0be99/extract/0.log" Dec 03 07:38:57 crc kubenswrapper[4818]: I1203 07:38:57.473137 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm_9897d8d1-c815-45ee-9c90-0b1e34a71340/util/0.log" Dec 03 07:38:57 crc kubenswrapper[4818]: I1203 07:38:57.640632 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm_9897d8d1-c815-45ee-9c90-0b1e34a71340/util/0.log" Dec 03 07:38:57 crc kubenswrapper[4818]: I1203 07:38:57.641695 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm_9897d8d1-c815-45ee-9c90-0b1e34a71340/pull/0.log" Dec 03 07:38:57 crc kubenswrapper[4818]: I1203 07:38:57.653669 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm_9897d8d1-c815-45ee-9c90-0b1e34a71340/pull/0.log" Dec 03 07:38:57 crc kubenswrapper[4818]: I1203 07:38:57.824124 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm_9897d8d1-c815-45ee-9c90-0b1e34a71340/util/0.log" Dec 03 07:38:57 crc kubenswrapper[4818]: I1203 07:38:57.857246 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm_9897d8d1-c815-45ee-9c90-0b1e34a71340/pull/0.log" Dec 03 07:38:57 crc kubenswrapper[4818]: I1203 07:38:57.860595 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83492hm_9897d8d1-c815-45ee-9c90-0b1e34a71340/extract/0.log" Dec 03 07:38:58 crc kubenswrapper[4818]: I1203 07:38:58.013278 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2kmpd_93370008-4229-4f0f-be59-8385406d7cd1/extract-utilities/0.log" Dec 03 07:38:58 crc kubenswrapper[4818]: I1203 07:38:58.153751 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2kmpd_93370008-4229-4f0f-be59-8385406d7cd1/extract-content/0.log" Dec 03 07:38:58 crc kubenswrapper[4818]: I1203 07:38:58.154134 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2kmpd_93370008-4229-4f0f-be59-8385406d7cd1/extract-utilities/0.log" Dec 03 07:38:58 crc kubenswrapper[4818]: I1203 07:38:58.156741 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2kmpd_93370008-4229-4f0f-be59-8385406d7cd1/extract-content/0.log" Dec 03 07:38:58 crc kubenswrapper[4818]: I1203 07:38:58.326473 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2kmpd_93370008-4229-4f0f-be59-8385406d7cd1/extract-utilities/0.log" Dec 03 07:38:58 crc kubenswrapper[4818]: I1203 07:38:58.343651 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2kmpd_93370008-4229-4f0f-be59-8385406d7cd1/extract-content/0.log" Dec 03 07:38:58 crc kubenswrapper[4818]: I1203 07:38:58.522510 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4wkmv_6bba00df-133e-484c-a16e-b2769ca31b7e/extract-utilities/0.log" Dec 03 07:38:58 crc kubenswrapper[4818]: I1203 07:38:58.748346 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4wkmv_6bba00df-133e-484c-a16e-b2769ca31b7e/extract-content/0.log" Dec 03 07:38:58 crc kubenswrapper[4818]: I1203 07:38:58.809342 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4wkmv_6bba00df-133e-484c-a16e-b2769ca31b7e/extract-utilities/0.log" Dec 03 07:38:58 crc kubenswrapper[4818]: I1203 07:38:58.826852 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4wkmv_6bba00df-133e-484c-a16e-b2769ca31b7e/extract-content/0.log" Dec 03 07:38:58 crc kubenswrapper[4818]: I1203 07:38:58.979672 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-2kmpd_93370008-4229-4f0f-be59-8385406d7cd1/registry-server/0.log" Dec 03 07:38:59 crc kubenswrapper[4818]: I1203 07:38:59.005164 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4wkmv_6bba00df-133e-484c-a16e-b2769ca31b7e/extract-content/0.log" Dec 03 07:38:59 crc kubenswrapper[4818]: I1203 07:38:59.023315 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4wkmv_6bba00df-133e-484c-a16e-b2769ca31b7e/extract-utilities/0.log" Dec 03 07:38:59 crc kubenswrapper[4818]: I1203 07:38:59.201017 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hl66m_a89a1dfa-1797-48f8-bd2d-f968c523e8df/marketplace-operator/0.log" Dec 03 07:38:59 crc kubenswrapper[4818]: I1203 07:38:59.220286 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4wkmv_6bba00df-133e-484c-a16e-b2769ca31b7e/registry-server/0.log" Dec 03 07:38:59 crc kubenswrapper[4818]: I1203 07:38:59.427351 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dm6xl_b2a86a74-f04c-49ed-a900-33a852b794a3/extract-utilities/0.log" Dec 03 07:38:59 crc kubenswrapper[4818]: I1203 07:38:59.574591 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dm6xl_b2a86a74-f04c-49ed-a900-33a852b794a3/extract-content/0.log" Dec 03 07:38:59 crc kubenswrapper[4818]: I1203 07:38:59.595179 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dm6xl_b2a86a74-f04c-49ed-a900-33a852b794a3/extract-utilities/0.log" Dec 03 07:38:59 crc kubenswrapper[4818]: I1203 07:38:59.598649 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dm6xl_b2a86a74-f04c-49ed-a900-33a852b794a3/extract-content/0.log" Dec 03 07:38:59 crc kubenswrapper[4818]: I1203 07:38:59.746197 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dm6xl_b2a86a74-f04c-49ed-a900-33a852b794a3/extract-utilities/0.log" Dec 03 07:38:59 crc kubenswrapper[4818]: I1203 07:38:59.747419 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dm6xl_b2a86a74-f04c-49ed-a900-33a852b794a3/extract-content/0.log" Dec 03 07:38:59 crc kubenswrapper[4818]: I1203 07:38:59.917558 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dm6xl_b2a86a74-f04c-49ed-a900-33a852b794a3/registry-server/0.log" Dec 03 07:38:59 crc kubenswrapper[4818]: I1203 07:38:59.949360 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hchtl_761a2fb4-d124-4db1-9924-0cd3d9546fa5/extract-utilities/0.log" Dec 03 07:39:00 crc kubenswrapper[4818]: I1203 07:39:00.101421 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hchtl_761a2fb4-d124-4db1-9924-0cd3d9546fa5/extract-content/0.log" Dec 03 07:39:00 crc kubenswrapper[4818]: I1203 07:39:00.158064 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hchtl_761a2fb4-d124-4db1-9924-0cd3d9546fa5/extract-utilities/0.log" Dec 03 07:39:00 crc kubenswrapper[4818]: I1203 07:39:00.159756 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hchtl_761a2fb4-d124-4db1-9924-0cd3d9546fa5/extract-content/0.log" Dec 03 07:39:00 crc kubenswrapper[4818]: I1203 07:39:00.306924 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hchtl_761a2fb4-d124-4db1-9924-0cd3d9546fa5/extract-utilities/0.log" Dec 03 07:39:00 crc kubenswrapper[4818]: I1203 07:39:00.313949 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hchtl_761a2fb4-d124-4db1-9924-0cd3d9546fa5/extract-content/0.log" Dec 03 07:39:01 crc kubenswrapper[4818]: I1203 07:39:01.043201 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hchtl_761a2fb4-d124-4db1-9924-0cd3d9546fa5/registry-server/0.log" Dec 03 07:39:13 crc kubenswrapper[4818]: I1203 07:39:13.301884 4818 patch_prober.go:28] interesting pod/machine-config-daemon-t9fzk container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 07:39:13 crc kubenswrapper[4818]: I1203 07:39:13.302468 4818 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 07:39:13 crc kubenswrapper[4818]: I1203 07:39:13.302519 4818 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" Dec 03 07:39:13 crc kubenswrapper[4818]: I1203 07:39:13.303130 4818 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595"} pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 07:39:13 crc kubenswrapper[4818]: I1203 07:39:13.303189 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerName="machine-config-daemon" containerID="cri-o://ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" gracePeriod=600 Dec 03 07:39:13 crc kubenswrapper[4818]: E1203 07:39:13.425952 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:39:13 crc kubenswrapper[4818]: I1203 07:39:13.662410 4818 generic.go:334] "Generic (PLEG): container finished" podID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" exitCode=0 Dec 03 07:39:13 crc kubenswrapper[4818]: I1203 07:39:13.662456 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerDied","Data":"ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595"} Dec 03 07:39:13 crc kubenswrapper[4818]: I1203 07:39:13.662494 4818 scope.go:117] "RemoveContainer" containerID="3696ecd804b3cc54d22f8baf9c3f030c9a0933db13e3f5ad4cac7c6ea9cfefef" Dec 03 07:39:13 crc kubenswrapper[4818]: I1203 07:39:13.663598 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:39:13 crc kubenswrapper[4818]: E1203 07:39:13.663974 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:39:24 crc kubenswrapper[4818]: I1203 07:39:24.741619 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:39:24 crc kubenswrapper[4818]: E1203 07:39:24.742465 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:39:36 crc kubenswrapper[4818]: I1203 07:39:36.738467 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:39:36 crc kubenswrapper[4818]: E1203 07:39:36.739262 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:39:49 crc kubenswrapper[4818]: I1203 07:39:49.738620 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:39:49 crc kubenswrapper[4818]: E1203 07:39:49.739419 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:40:04 crc kubenswrapper[4818]: I1203 07:40:04.738414 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:40:04 crc kubenswrapper[4818]: E1203 07:40:04.739079 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:40:18 crc kubenswrapper[4818]: I1203 07:40:18.739080 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:40:18 crc kubenswrapper[4818]: E1203 07:40:18.740392 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:40:31 crc kubenswrapper[4818]: I1203 07:40:31.738279 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:40:31 crc kubenswrapper[4818]: E1203 07:40:31.739113 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:40:45 crc kubenswrapper[4818]: I1203 07:40:45.565639 4818 generic.go:334] "Generic (PLEG): container finished" podID="af139d2b-a174-40c0-b8ae-d6db7c004db3" containerID="c2d67076699bf0d1b5c53aebd9b2abfeca7a5bcbbbee3e1ff2306cc1d3f5aaa0" exitCode=0 Dec 03 07:40:45 crc kubenswrapper[4818]: I1203 07:40:45.565836 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-thj5g/must-gather-9qkqz" event={"ID":"af139d2b-a174-40c0-b8ae-d6db7c004db3","Type":"ContainerDied","Data":"c2d67076699bf0d1b5c53aebd9b2abfeca7a5bcbbbee3e1ff2306cc1d3f5aaa0"} Dec 03 07:40:45 crc kubenswrapper[4818]: I1203 07:40:45.566843 4818 scope.go:117] "RemoveContainer" containerID="c2d67076699bf0d1b5c53aebd9b2abfeca7a5bcbbbee3e1ff2306cc1d3f5aaa0" Dec 03 07:40:46 crc kubenswrapper[4818]: I1203 07:40:46.458345 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-thj5g_must-gather-9qkqz_af139d2b-a174-40c0-b8ae-d6db7c004db3/gather/0.log" Dec 03 07:40:46 crc kubenswrapper[4818]: I1203 07:40:46.739284 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:40:46 crc kubenswrapper[4818]: E1203 07:40:46.739756 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:40:56 crc kubenswrapper[4818]: I1203 07:40:56.898937 4818 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-thj5g/must-gather-9qkqz"] Dec 03 07:40:56 crc kubenswrapper[4818]: I1203 07:40:56.899613 4818 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-thj5g/must-gather-9qkqz" podUID="af139d2b-a174-40c0-b8ae-d6db7c004db3" containerName="copy" containerID="cri-o://1d3e56afafdfed16684e2932255bdbb3c45bb35aba8a20339aaf2d3a2d2ff0b1" gracePeriod=2 Dec 03 07:40:56 crc kubenswrapper[4818]: I1203 07:40:56.908038 4818 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-thj5g/must-gather-9qkqz"] Dec 03 07:40:57 crc kubenswrapper[4818]: I1203 07:40:57.683115 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-thj5g_must-gather-9qkqz_af139d2b-a174-40c0-b8ae-d6db7c004db3/copy/0.log" Dec 03 07:40:57 crc kubenswrapper[4818]: I1203 07:40:57.683741 4818 generic.go:334] "Generic (PLEG): container finished" podID="af139d2b-a174-40c0-b8ae-d6db7c004db3" containerID="1d3e56afafdfed16684e2932255bdbb3c45bb35aba8a20339aaf2d3a2d2ff0b1" exitCode=143 Dec 03 07:40:57 crc kubenswrapper[4818]: I1203 07:40:57.683779 4818 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d41ea4397ee8fe6b1080387c5816db86b4092454b715fc376677ac73dac9e7b" Dec 03 07:40:57 crc kubenswrapper[4818]: I1203 07:40:57.738041 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:40:57 crc kubenswrapper[4818]: E1203 07:40:57.738535 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:40:57 crc kubenswrapper[4818]: I1203 07:40:57.836101 4818 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-thj5g_must-gather-9qkqz_af139d2b-a174-40c0-b8ae-d6db7c004db3/copy/0.log" Dec 03 07:40:57 crc kubenswrapper[4818]: I1203 07:40:57.836642 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-thj5g/must-gather-9qkqz" Dec 03 07:40:57 crc kubenswrapper[4818]: I1203 07:40:57.867780 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtt4z\" (UniqueName: \"kubernetes.io/projected/af139d2b-a174-40c0-b8ae-d6db7c004db3-kube-api-access-gtt4z\") pod \"af139d2b-a174-40c0-b8ae-d6db7c004db3\" (UID: \"af139d2b-a174-40c0-b8ae-d6db7c004db3\") " Dec 03 07:40:57 crc kubenswrapper[4818]: I1203 07:40:57.868131 4818 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/af139d2b-a174-40c0-b8ae-d6db7c004db3-must-gather-output\") pod \"af139d2b-a174-40c0-b8ae-d6db7c004db3\" (UID: \"af139d2b-a174-40c0-b8ae-d6db7c004db3\") " Dec 03 07:40:57 crc kubenswrapper[4818]: I1203 07:40:57.873646 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af139d2b-a174-40c0-b8ae-d6db7c004db3-kube-api-access-gtt4z" (OuterVolumeSpecName: "kube-api-access-gtt4z") pod "af139d2b-a174-40c0-b8ae-d6db7c004db3" (UID: "af139d2b-a174-40c0-b8ae-d6db7c004db3"). InnerVolumeSpecName "kube-api-access-gtt4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 07:40:57 crc kubenswrapper[4818]: I1203 07:40:57.969431 4818 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtt4z\" (UniqueName: \"kubernetes.io/projected/af139d2b-a174-40c0-b8ae-d6db7c004db3-kube-api-access-gtt4z\") on node \"crc\" DevicePath \"\"" Dec 03 07:40:58 crc kubenswrapper[4818]: I1203 07:40:58.037378 4818 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af139d2b-a174-40c0-b8ae-d6db7c004db3-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "af139d2b-a174-40c0-b8ae-d6db7c004db3" (UID: "af139d2b-a174-40c0-b8ae-d6db7c004db3"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 07:40:58 crc kubenswrapper[4818]: I1203 07:40:58.070986 4818 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/af139d2b-a174-40c0-b8ae-d6db7c004db3-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 07:40:58 crc kubenswrapper[4818]: I1203 07:40:58.696436 4818 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-thj5g/must-gather-9qkqz" Dec 03 07:40:58 crc kubenswrapper[4818]: I1203 07:40:58.774637 4818 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af139d2b-a174-40c0-b8ae-d6db7c004db3" path="/var/lib/kubelet/pods/af139d2b-a174-40c0-b8ae-d6db7c004db3/volumes" Dec 03 07:41:12 crc kubenswrapper[4818]: I1203 07:41:12.745890 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:41:12 crc kubenswrapper[4818]: E1203 07:41:12.748416 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:41:23 crc kubenswrapper[4818]: I1203 07:41:23.737927 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:41:23 crc kubenswrapper[4818]: E1203 07:41:23.739508 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:41:37 crc kubenswrapper[4818]: I1203 07:41:37.739160 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:41:37 crc kubenswrapper[4818]: E1203 07:41:37.739967 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:41:51 crc kubenswrapper[4818]: I1203 07:41:51.738357 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:41:51 crc kubenswrapper[4818]: E1203 07:41:51.739631 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:41:58 crc kubenswrapper[4818]: I1203 07:41:58.856733 4818 scope.go:117] "RemoveContainer" containerID="1d3e56afafdfed16684e2932255bdbb3c45bb35aba8a20339aaf2d3a2d2ff0b1" Dec 03 07:41:58 crc kubenswrapper[4818]: I1203 07:41:58.883339 4818 scope.go:117] "RemoveContainer" containerID="c2d67076699bf0d1b5c53aebd9b2abfeca7a5bcbbbee3e1ff2306cc1d3f5aaa0" Dec 03 07:42:04 crc kubenswrapper[4818]: I1203 07:42:04.738872 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:42:04 crc kubenswrapper[4818]: E1203 07:42:04.739887 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:42:16 crc kubenswrapper[4818]: I1203 07:42:16.738022 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:42:16 crc kubenswrapper[4818]: E1203 07:42:16.739142 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:42:28 crc kubenswrapper[4818]: I1203 07:42:28.738586 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:42:28 crc kubenswrapper[4818]: E1203 07:42:28.739494 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:42:41 crc kubenswrapper[4818]: I1203 07:42:41.738017 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:42:41 crc kubenswrapper[4818]: E1203 07:42:41.738778 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:42:55 crc kubenswrapper[4818]: I1203 07:42:55.737683 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:42:55 crc kubenswrapper[4818]: E1203 07:42:55.738916 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:43:07 crc kubenswrapper[4818]: I1203 07:43:07.788937 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:43:07 crc kubenswrapper[4818]: E1203 07:43:07.789705 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:43:19 crc kubenswrapper[4818]: I1203 07:43:19.738027 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:43:19 crc kubenswrapper[4818]: E1203 07:43:19.738728 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:43:32 crc kubenswrapper[4818]: I1203 07:43:32.745602 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:43:32 crc kubenswrapper[4818]: E1203 07:43:32.746461 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:43:45 crc kubenswrapper[4818]: I1203 07:43:45.738139 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:43:45 crc kubenswrapper[4818]: E1203 07:43:45.739276 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:44:00 crc kubenswrapper[4818]: I1203 07:44:00.738409 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:44:00 crc kubenswrapper[4818]: E1203 07:44:00.739404 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:44:11 crc kubenswrapper[4818]: I1203 07:44:11.739073 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:44:11 crc kubenswrapper[4818]: E1203 07:44:11.740030 4818 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t9fzk_openshift-machine-config-operator(dd398f0d-786f-44e7-a64b-7cf7516d5d54)\"" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" podUID="dd398f0d-786f-44e7-a64b-7cf7516d5d54" Dec 03 07:44:22 crc kubenswrapper[4818]: I1203 07:44:22.745669 4818 scope.go:117] "RemoveContainer" containerID="ca67d8f44ebe95e36883e30f697d9a5b38875ce7b987b798f3b8b30c695f2595" Dec 03 07:44:23 crc kubenswrapper[4818]: I1203 07:44:23.965619 4818 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t9fzk" event={"ID":"dd398f0d-786f-44e7-a64b-7cf7516d5d54","Type":"ContainerStarted","Data":"ea3906dc68331b06f07bcc552e523c403515490f666198940d4b64ae2d3f861b"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515113765151024452 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015113765152017370 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015113753615016514 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015113753616015465 5ustar corecore